1. 程式人生 > >hadoop0.20.2偽分布式環境搭建

hadoop0.20.2偽分布式環境搭建

VM cal local 下載 lld map same ces task

雖然現在hadoop版本已經到了主流2點多的時代,但是對於學習大數據而言,我還是選擇從以前老的版本0.20.2學起。

下面就是偽分布式的環境搭建過程。


hadoop下載地址:

http://archive.apache.org/dist/hadoop/core/hadoop-0.20.2/hadoop-0.20.2.tar.gz


linux系統版本:centos7


1、配置主機名

[root@localhost ~]# vi /etc/sysconfig/network

# Created by anaconda

master1


[root@localhost ~]# hostname master1


2、創建管理hadoop的組和用戶

[root@master1 ~]# groupadd hduser


[root@master1 ~]# useradd -g hduser hduser


[root@master1 ~]# passwd hduser


3、hosts主機名ip解析

[root@master1 ~]# vi /etc/hosts

192.168.11.131 master1


4、配置hadoop的sudoers權限

[root@master1 ~]# vi /etc/sudoers

hduser ALL=(ALL) NOPASSWD:ALL


5、關閉selinux和防火墻

[root@master1 ~]# vi /etc/sysconfig/selinux

SELINUX=enforcing --> SELINUX=disabled


[root@master1 ~]# systemctl stop firewalld

[root@master1 ~]# systemctl disable firewalld


6、解壓包

[root@master1 ~]# su hduser

[hduser@master1 root]$ cd


[hduser@master1 ~]$ ll *tar*

-rw-r--r--. 1 root root 44575568 Jun 16 17:24 hadoop-0.20.2.tar.gz

-rw-r--r--. 1 root root 288430080 Mar 16 2016 jdk1.7.0_79.tar


[hduser@master1 ~]$ tar xf jdk1.7.0_79.tar


[hduser@master1 ~]$ tar zxf hadoop-0.20.2.tar.gz


[hduser@master1 ~]$ mv jdk1.7.0_79 jdk


[hduser@master1 ~]$ mv hadoop-0.20.2 hadoop


7、配置java環境

[hduser@master1 ~]$ vi .bashrc

export JAVA_HOME=/home/hduser/jdk

export JRE_HOME=$JAVA_HOME/jre

export PATH=$PATH:$JAVA_HOME/bin

export CLASSPATH=./:$JAVA_HOME/lib:$JAVA_HOME/jre/lib


[hduser@master1 ~]$ source .bashrc


[hduser@master1 ~]$ java -version

java version "1.7.0_79"

Java(TM) SE Runtime Environment (build 1.7.0_79-b15)

Java HotSpot(TM) 64-Bit Server VM (build 24.79-b02, mixed mode)


8、配置hadoop

[hduser@master1 conf]$ pwd

/home/hduser/hadoop/conf

[hduser@master1 conf]$ vi hadoop-env.sh

export JAVA_HOME=/home/hduser/jdk


[hduser@master1 conf]$ vi core-site.xml

<configuration>

<property>

<name>fs.default.name</name>

<value>hdfs://master1:9000</value>

</property>

</configuration>


[hduser@master1 conf]$ sudo mkdir -p /data/hadoop/data

[hduser@master1 conf]$ sudo chown -R hduser:hduser /data/hadoop/data


[hduser@master1 conf]$ vi hdfs-site.xml

<configuration>

<property>

<name>dfs.data.dir</name>

<value>/data/hadoop/data</value>

</property>

<property>

<name>dfs.replication</name>

<value>1</value>

</property>

</configuration>


[hduser@master1 conf]$ vi mapred-site.xml

<configuration>

<property>

<name>mapred.job.tracker</name>

<value>master1:9001</value>

</property>

</configuration>


9、做無密碼認證

[hduser@master1 conf]$ cd

[hduser@master1 ~]$ ssh-keygen -t rsa

Generating public/private rsa key pair.

Enter file in which to save the key (/home/hduser/.ssh/id_rsa):

Created directory '/home/hduser/.ssh'.

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/hduser/.ssh/id_rsa.

Your public key has been saved in /home/hduser/.ssh/id_rsa.pub.

The key fingerprint is:

SHA256:qRJhPSF32QDs9tU3e0/mAx/EBC2MHamGv2WPvUw19/M hduser@master1

The key's randomart image is:

+---[RSA 2048]----+

| ..+.o+ +o= |

| +.o. .. = o |

| o.o ... + |

| . .o. o.o. oo |

| .. .S.o ..+o|

| . .. . +..O|

| . . + *B+|

| . . .o==|

| oE|

+----[SHA256]-----+


一路enter鍵


[hduser@master1 ~]$ cd .ssh

[hduser@master1 .ssh]$ ls

id_rsa id_rsa.pub


[hduser@master1 .ssh]$ cp id_rsa.pub authorized_keys


10、格式化文件系統

[hduser@master1 .ssh]$ cd

[hduser@master1 ~]$ cd hadoop/bin

[hduser@master1 bin]$ ./hadoop namenode -format

18/06/19 04:02:12 INFO namenode.NameNode: STARTUP_MSG:

/************************************************************

STARTUP_MSG: Starting NameNode

STARTUP_MSG: host = master1/192.168.11.131

STARTUP_MSG: args = [-format]

STARTUP_MSG: version = 0.20.2

STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010

************************************************************/

18/06/19 04:02:13 INFO namenode.FSNamesystem: fsOwner=hduser,hduser

18/06/19 04:02:13 INFO namenode.FSNamesystem: supergroup=supergroup

18/06/19 04:02:13 INFO namenode.FSNamesystem: isPermissionEnabled=true

18/06/19 04:02:13 INFO common.Storage: Image file of size 96 saved in 0 seconds.

18/06/19 04:02:13 INFO common.Storage: Storage directory /tmp/hadoop-hduser/dfs/name has been successfully formatted.

18/06/19 04:02:13 INFO namenode.NameNode: SHUTDOWN_MSG:

/************************************************************

SHUTDOWN_MSG: Shutting down NameNode at master1/192.168.11.131

************************************************************/


11、啟動服務

[hduser@master1 bin]$ ./start-all.sh

starting namenode, logging to /home/hduser/hadoop/bin/../logs/hadoop-hduser-namenode-master1.out

The authenticity of host 'localhost (::1)' can't be established.

ECDSA key fingerprint is SHA256:OXYl4X6F6g4TV7YriZaSvuBIFM840h/qTg8/B7BUil0.

ECDSA key fingerprint is MD5:b6:b6:04:2d:49:70:8b:ed:65:00:e2:05:b0:95:5b:6d.

Are you sure you want to continue connecting (yes/no)? yes

localhost: Warning: Permanently added 'localhost' (ECDSA) to the list of known hosts.

localhost: starting datanode, logging to /home/hduser/hadoop/bin/../logs/hadoop-hduser-datanode-master1.out

localhost: starting secondarynamenode, logging to /home/hduser/hadoop/bin/../logs/hadoop-hduser-secondarynamenode-master1.out

starting jobtracker, logging to /home/hduser/hadoop/bin/../logs/hadoop-hduser-jobtracker-master1.out

localhost: starting tasktracker, logging to /home/hduser/hadoop/bin/../logs/hadoop-hduser-tasktracker-master1.out


12、查看服務

[hduser@master1 bin]$ jps

1867 JobTracker

1804 SecondaryNameNode

1597 NameNode

1971 TaskTracker

2011 Jps

1710 DataNode

[hduser@master1 bin]$


13、瀏覽器查看服務狀態

使用web查看HSFS運行狀態

在瀏覽器輸入

http://192.168.11.131:50030

使用web查看MapReduce運行狀態

在瀏覽器輸入

http://192.168.11.131:50070


hadoop0.20.2偽分布式環境搭建