export JAVA_HOME=/usr/local/jdk1.8.0_291 export PATH=${JAVA_HOME}/bin:${PATH} export M2_HOME=/usr/local/maven export PATH=${M2_HOME}/bin:${PATH} export SCALA_HOME=/usr/local/scala export PATH=$PATH:$SCALA_HOME/bin export KAFKA_HOME=/usr/local/bigdata/kafka export PATH=$PATH:$KAFKA_HOME/bin export HADOOP_HOME=/usr/local/bigdata/hadoop export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin export SPARK_HOME=/usr/local/spark export PATH=$PATH:$SPARK_HOME/bin:$SPARK_HOME/sbin export ZK_HOME=/usr/local/bigdata/zookeeper export PATH=$PATH:$ZK_HOME/bin export Hbase_HOME=/usr/local/bigdata/hbase export PATH=$PATH:$Hbase_HOME/bin配置JDK和Maven
版本: JDK: 1.8 Maven: 1.8 配置环境变量: export JAVA_HOME=/usr/local/jdk1.8.0_291 export PATH=${JAVA_HOME}/bin:${PATH} export M2_HOME=/usr/local/maven export PATH=${M2_HOME}/bin:${PATH}Hadoop 版本 hadoop-2.10.1
// 进入hadoop安装目录 sudo vim etc/hadoop/hadoop-env.sh // 设置JAVA_HOME 参数 根据自身电脑的Java位置 export JAVA_HOME=/usr/local/jdk1.8.0_291 //保存退出
//配置 etc/hadoop/core-site.xmlhadoop.tmp.dir file:/usr/local/hadoop/tmp Abase for other temporary directories. fs.defaultFS hdfs://localhost:9000
// 配置etc/hadoop/hdfs-site.xmldfs.replication 1 dfs.namenode.name.dir file:/usr/local/hadoop/tmp/dfs/name dfs.datanode.data.dir file:/usr/local/hadoop/tmp/dfs/data
//配置ssh免密码登录 ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys chmod 0600 ~/.ssh/authorized_keys //格式化文件系统 不要重复格式化,会导致Datanode节点启动不了 bin/hdfs namenode -format // 启动hadoop服务 sbin/start-dfs.sh
// 如果出现 The authenticity of host 'master (192.168.5.133)' can‘t be established. #打开文件 vi /etc/ssh/ssh_config #在最后面添加如下语句即可 StrictHostKeyChecking no UserKnownHostsFile /dev/nullzookeeper
//配置zookeeper //创建data 存放数据 mkdir data //编辑配置文件 sudo mv conf/zoo_sample.cfg conf/zoo.cfg sudo vim conf/zoo.cfg //配置如下 # The number of milliseconds of each tick tickTime=2000 # The number of ticks that the initial # synchronization phase can take initLimit=10 # The number of ticks that can pass between # sending a request and getting an acknowledgement syncLimit=5 # the directory where the snapshot is stored. # do not use /tmp for storage, /tmp here is just # example sakes. dataDir=/usr/local/bigdata/zookeeper/data # the port at which the clients will connect clientPort=2181 //启动zookeeper sudo bash bin/zkServer.sh start //failed to start mvn package -Dmaven.test.skip=trueHbase
# 下载:https://hbase.apache.org/downloads.html sudo tar -vxzf hbase-2.4.2-bin.tar.gz cp -a hbase-2.4.2 /usr/local/hbase-2.4.2 # 配置hbase-env.sh sudo vim conf/hbase-env.sh //设置如下参数 export JAVA_HOME=/usr/local/jdk1.8.0_291 export Hbase_MANAGES_ZK=false
// 配置habse-site.xml sudo vim conf/hbase-site.xml //配置如下Kafkahbase.rootdir hdfs://hadoop:9000/hbase hbase.cluster.distributed true hbase.zookeeper.property.clientPort 2181 hbase.zookeeper.quorum hadoop # 启动Hbase ./bin/start-hbase.sh hbase.zookeeper.property.dataDir /usr/local/bigdata/zookeeper/zoo_data
# 配置kafka # sudo vim config/server.properties broker.id=0 port=9092 log.dirs=/usr/local/bigdata/kafka/kafkalog # 启动Kafka sudo bash bin/kafka-server-start.sh config/server.properties
欢迎分享,转载请注明来源:内存溢出
评论列表(0条)