Hadoop安装(3.3.1)

Hadoop安装(3.3.1),第1张

Hadoop安装(3.3.1) Hadoop安装(3.3.1)

1.设置主机名

hostname hadoop01 #设置临时主机名
vi /etc/hostname #设置永久主机名

2.设置hosts

vi /etc/hosts

192.168.136.103 hadoop01
192.168.136.104 hadoop02
192.168.136.105 hadoop03

3.关闭防火墙

systemctl stop firewalld #关闭防火墙
systemctl disable firewalld #禁用防火墙

4.ssh免密登陆

#在主节点上执行
ssh-keygen -t rsa #输入内容直接回车
#将生成的公钥添加到免密登陆keys中
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
#将公钥复制到2台从节点上
scp ~/.ssh/id_rsa.pub root@hadoop02:~/
scp ~/.ssh/id_rsa.pub root@hadoop03:~/
#在从节点上将主节点的公钥添加到免密登陆keys中
cat ~/id_rsa.pub >> ~/.ssh/authorized_keys

5.ntpdate时间同步

yum install ntpdate
#同步时间
ntpdate -u ntp.sjtu.edu.cn
#定时同步,每小时整点执行一次
vi /etc/crontab
0 */1 * * * root ntpdate -u ntp.sjtu.edu.cn

6.下载hadoop

#下载
wget https://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/stable/hadoop-3.3.1.tar.gz
#解压
tar -zxvf hadoop-3.3.1.tar.gz

7.编辑etc/hadoop/hadoop-env.sh

cd hadoop-3.3.1/etc/hadoop/
vi hadoop-env.sh
#添加如下内容
export JAVA_HOME=/home/jdk1.8/jdk1.8.0_191
export HADOOP_HOME=/home/hadoop3.3.1/hadoop-3.3.1
export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop
export HADOOP_LOG_DIR=${HADOOP_HOME}/logs
export HADOOP_PID_DIR=${HADOOP_HOME}/pid

8.编辑etc/hadoop/core-site.xml


    
        fs.defaultFS
        hdfs://hadoop01:9000
    
    
        hadoop.tmp.dir
        /home/hadoop_repo
   

9.编辑etc/hadoop/hdfs-site.xml


    
        dfs.replication
        2
    
    
        dfs.namenode.secondary.http-address
        hadoop01:9868
    

10.编辑etc/hadoop/mapred-site.xml


    
        mapreduce.framework.name
        yarn
    

11.编辑etc/hadoop/yarn-site.xml


    
        yarn.nodemanager.aux-services
        mapreduce_shuffle
    
    
        yarn.nodemanager.env-whitelist
        JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME,HADOOP_MAPRED_HOME
    
	
		yarn.resourcemanager.hostname
		hadoop01
	

12.编辑etc/hadoop/workers

hadoop02
hadoop03

13.编辑启动脚本

cd sbin

vi start-dfs.sh
HDFS_DATANODE_USER=root
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root
HDFS_SECONDARYNAMENODE_USER=root

vi stop-dfs.sh
HDFS_DATANODE_USER=root
HDFS_DATANODE_SECURE_USER=hdfs
HDFS_NAMENODE_USER=root
HDFS_SECONDARYNAMENODE_USER=root

vi start-yarn.sh
YARN_RESOURCEMANAGER_USER=root
HADOOP_SECURE_DN_USER=yarn
YARN_NODEMANAGER_USER=root

vi stop-yarn.sh
YARN_RESOURCEMANAGER_USER=root
HADOOP_SECURE_DN_USER=yarn
YARN_NODEMANAGER_USER=root

14.格式化HDFS

#格式化namenode节点
bash hdfs namenode -format
#出现如下信息说明格式化成功
common.Storage: Storage directory /home/hadoop_repo/dfs/name has been successfully formatted.

15.启动集群

sbin/start-all.sh #启动集群
sbin/stop-all.sh #停止集群

16.验证

#主节点jps
NameNode
SecondaryNameNode
ResourceManager
#从节点jps
DataNode
NodeManager

17.查看集群信息

#资源信息查看
http://192.168.136.103:8088/
#存储节点信息查看
http://192.168.136.103:9870/

欢迎分享,转载请注明来源:内存溢出

原文地址: https://outofmemory.cn/zaji/5652597.html

(0)
打赏 微信扫一扫 微信扫一扫 支付宝扫一扫 支付宝扫一扫
上一篇 2022-12-16
下一篇 2022-12-16

发表评论

登录后才能评论

评论列表(0条)

保存