1.系统初始化设置

sudo apt-get install ssh
sudo apt-get install pdshvi /etc/hosts
172.105.193.146 master
172.105.195.123 slave1
#172.105.193.146 slave2vi /etc/hostname
修改为对应主机名cd /root
ssh-keygen -t rsassh-copy-id -i /root/.ssh/id_rsa.pub root@master
ssh-copy-id -i /root/.ssh/id_rsa.pub root@slave1
#ssh-copy-id -i /root/.ssh/id_rsa.pub root@slave2

2.检测证书登录是否正常

ssh root@master
ssh root@slave1
#ssh root@slave2

3.解压文件

tar -xvzf jdk-8u271-linux-x64.tar.gz

4.移动文件

mv hadoop-3.3.0 /usr/local/hadoop
mv jdk1.8.0_271 /usr/local/java

5.配置环境变量

vi /etc/profile
export JAVA_HOME=/usr/local/java
export HADOOP_HOME=/usr/local/hadoop
export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
export PDSH_RCMD_TYPE=ssh

6.使环境变量生效

source /etc/profile

7.切换至hadoop配置文件目录

cd /usr/local/hadoop/etc/hadoop

8.修改配置文件

vi /usr/local/hadoop/etc/hadoop/core-site.xml
<configuration>
    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://master:9000</value>
    </property>
</configuration>

vi /usr/local/hadoop/etc/hadoop/hdfs-site.xml
<configuration>
        <property>
                <name>dfs.namenode.http-address</name>
                <value>master:50070</value>
        </property>
        <property>
                <name>dfs.namenode.secondary.http-address</name>
                <value>slave1:50090</value>
        </property>
        <property>
                <name>dfs.namenode.name.dir</name>
                <value>/usr/local/hadoop/hadoopname</value>
        </property>
        <property>
                <name>dfs.replication</name>
                <value>2</value>
         </property>
         <property>
                <name>dfs.datanode.data.dir</name>
                <value>/usr/local/hadoop/hadoopdata</value>
        </property>
</configuration>

vi /usr/local/hadoop/etc/hadoop/mapred-site.xml
<configuration>
     <property>
          <name>mapreduce.framework.name</name>
          <value>yarn</value>
     </property>
     <property>
          <name>mapreduce.application.classpath</name>
          <value>
           /usr/local/hadoop/etc/hadoop,
           /usr/local/hadoop/share/hadoop/common/*,
           /usr/local/hadoop/share/hadoop/common/lib/*,
           /usr/local/hadoop/share/hadoop/hdfs/*,
           /usr/local/hadoop/share/hadoop/hdfs/lib/*,
           /usr/local/hadoop/share/hadoop/mapreduce/*,
           /usr/local/hadoop/share/hadoop/mapreduce/lib/*,
           /usr/local/hadoop/share/hadoop/yarn/*,
           /usr/local/hadoop/share/hadoop/yarn/lib/*
          </value>
     </property>
</configuration>

vi /usr/local/hadoop/etc/hadoop/workers
master
slave1
slave2

vi /usr/local/hadoop/etc/hadoop/hadoop-env.sh
export JAVA_HOME=/usr/local/java
export HADOOP_HOME=/usr/local/hadoop

vi /usr/local/hadoop/etc/hadoop/yarn-env.sh
export JAVA_HOME=/usr/local/java

vi /usr/local/hadoop/etc/hadoop/yarn-site.xml
<configuration>
    <property>
         <name>yarn.resourcemanager.hostname</name>
         <value>master</value>
    </property>
    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
    <property>
        <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
        <value>org.apache.hadoop.mapred.ShuffleHandler</value>
    </property>
</configuration>

vi /usr/local/hadoop/sbin/start-all.sh
HDFS_DATANODE_USER=root 
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root 
HDFS_SECONDARYNAMENODE_USER=root
YARN_RESOURCEMANAGER_USER=root 
HADOOP_SECURE_DN_USER=yarn 
YARN_NODEMANAGER_USER=root

vi /usr/local/hadoop/sbin/stop-all.sh
HDFS_DATANODE_USER=root 
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root 
HDFS_SECONDARYNAMENODE_USER=root
YARN_RESOURCEMANAGER_USER=root 
HADOOP_SECURE_DN_USER=yarn 
YARN_NODEMANAGER_USER=root

vi /usr/local/hadoop/sbin/start-dfs.sh
HDFS_DATANODE_USER=root 
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root 
HDFS_SECONDARYNAMENODE_USER=root

vi /usr/local/hadoop/sbin/stop-dfs.sh
HDFS_DATANODE_USER=root 
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root 
HDFS_SECONDARYNAMENODE_USER=root

vi /usr/local/hadoop/sbin/stop-yarn.sh
YARN_RESOURCEMANAGER_USER=root 
HADOOP_SECURE_DN_USER=yarn 
YARN_NODEMANAGER_USER=root


vi /usr/local/hadoop/sbin/start-yarn.sh
YARN_RESOURCEMANAGER_USER=root 
HADOOP_SECURE_DN_USER=yarn 
YARN_NODEMANAGER_USER=root

9.将配置好的环境分配至各个节点

scp -r /usr/local/hadoop slave1:/usr/local
#scp -r /usr/local/hadoop slave2:/usr/local

scp -r /usr/local/java slave1:/usr/local
#scp -r /usr/local/java slave2:/usr/local

scp -r /etc/profile slave1:/etc/
#scp -r /etc/profile slave2:/etc/

10.

mkdir /usr/local/hadoop/hadoopname

11.

source /etc/profile
hdfs namenode -format
#hdfs datanode -format
Last modification:October 23, 2020
如果觉得我的文章对你有用,请随意赞赏