1. 程式人生 > >cdh5.7.0偽分散式叢集之hadoop安裝

cdh5.7.0偽分散式叢集之hadoop安裝

基本環境及軟體:

軟體版本 軟體包
centos-6.4  
JDK-1.8 jdk-8u191-linux-x64.tar.gz
hadoop-2.6.0 hadoop-2.6.0-cdh5.7.0.tar.gz

軟體安裝包官網下載地址 :http://archive-primary.cloudera.com/cdh5/cdh/5/ 

設定免密碼登入 

1.生成本機的公鑰,過程中不斷敲回車即可,ssh-keygen命令預設會將公鑰放在/root/.ssh目錄下

# ssh-keygen -t rsa

2.將公鑰複製為authorized_keys檔案,此時使用ssh連線本機就不需要輸入密碼了

# cd /root/.ssh

# cp id_rsa.pub authorized_keys

3.設定ip和hostname的對應關係

# vi /etc/hosts 

192.168.0.104 bigdata

4.配置是否配置成功

# ssh bigdata 

安裝JDK1.8

1.在虛擬機器中的建立app目錄,存放安裝的軟體

# cd /usr/local/

# mkdir app

2.將下載好的jdk-8u191-linux-x64.tar.gz安裝包上傳到虛擬機器的 /use/local/app/ 目錄下

3.對安裝包進行解壓縮

# tar -zxvf jdk-8u191-linux-x64.tar.gz

4.配置環境變數,

# vi /etc/profile


#set java environment
export JAVA_HOME=/usr/local/app/jdk1.8.0_191  
export JRE_HOME=$JAVA_HOME/jre  
export PATH=$PATH:$JAVA_HOME/bin  
export CLASSPATH=./:$JAVA_HOME/lib:$JAVA_HOME/jre/lib

5.使環境變數生效

# source /etc/profile 

6.驗證是否安裝成功

 #  java -version

 # java

# javac

 

 

 

安裝hadoop

1、將下載好的hadoop-2.6.0-cdh5.7.0.tar.gz,上傳到虛擬機器的/usr/local/app目錄下。(http://archive.cloudera.com/cdh5/cdh/5/)
2、將hadoop包進行解壓縮:

# tar -zxvf hadoop-2.6.0-cdh5.7.0.tar.gz

3、對hadoop目錄進行重新命名:

# mv hadoop-2.6.0-cdh5.7.0 hadoop-2.6.0

4、配置hadoop相關環境變數

# vi ~/.bashrc


#set hadoop environment
export HADOOP_HOME=/usr/local/app/hadoop-2.6.0
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin


# source ~/.bashrc

5、建立/usr/local/app/data目錄 ,並且進入到/usr/local/app/hadoop-2.6.0/etc/hadoop 目錄下

# mkdir /usr/local/app/data

# cd /usr/local/app/hadoop-2.6.0/etc/hadoop

6.修改hadoop-env.sh

export JAVA_HOME=/usr/local/app/jdk1.8.0_191
export HADOOP_PID_DIR=/usr/local/app/data/tmp


7.配置mapred-env.sh

export HADOOP_MAPRED_PID_DIR=/usr/local/app/data/tmp

8.配置yarn-env.sh

export YARN_PID_DIR=/usr/local/app/data/tmp

9.配置core-site.xml

<property>
  <name>fs.default.name</name>
  <value>hdfs://bigdata:9000</value>
</property>

10.配置hdfs-site.xml

<property>
  <name>dfs.name.dir</name>
  <value>/usr/local/app/data/namenode</value>
</property>
<property>
  <name>dfs.data.dir</name>
  <value>/usr/local/app/data/datanode</value>
</property>
<property>
  <name>dfs.tmp.dir</name>
  <value>/usr/local/app/data/tmp</value>
</property>
<property>
  <name>dfs.replication</name>
  <value>1</value>
</property>

11.配置mapred-site.xml,

# cp mapred-site.xml.template mapred-site.xml

# vim mapred-site.xml

<property>
  <name>mapreduce.framework.name</name>
  <value>yarn</value>
</property>

12.配置yarn-site.xml

<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>bigdata</value>
</property>
<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>

啟動hdfs叢集


1、格式化namenode:

# hdfs namenode -format

2、啟動hdfs叢集:

# start-dfs.sh

3、驗證啟動是否成功:

a. jps驗證:namenode、datanode、secondarynamenode


b.50070驗證埠:http://192.168.0.104:50070

c、讀寫操作測試驗證:

# vi hello.txt

       hadoop hive spark

# hdfs dfs -mkdir /test
# hdfs dfs -put hello.txt /test/hello.txt
# hdfs dfs -text /test/hello.txt

啟動yarn叢集
1、啟動yarn叢集:

# start-yarn.sh

2、驗證啟動是否成功:
a.jps驗證:resourcemanager、nodemanager


b.8088埠驗證:http://192.168.0.104:8088

 

到此Hadoop偽分散式叢集安裝完成,有錯誤之處,請多多指教