1. 程式人生 > >hadoop 集群安裝

hadoop 集群安裝

reduce clas export mapr cond res oot should etc

一、安裝集群

A、上傳HADOOP安裝包

B、規劃安裝目錄 /export/servers/hadoop-2.8.4

C、解壓安裝包

D、修改配置文件 $HADOOP_HOME/etc/hadoop/

1、hadoop-env.sh

  export JAVA_HOME=/export/servers/jdk1.8.0_11

2、core-site.xml  

  <configuration>

    <property>

    <name>fs.defaultFS</name>

    <value>hdfs://hadoop1:9000</value>

    </property>

    <property>

    <name>hadoop.tmp.dir</name>

    <value>/export/servers/hadoop/tmp</value>

    </property>

  </configuration>

3、hdfs-site.xml

<configuration>
<property>
<name>dfs.name.dir</name>
<value>/export/servers/hadoop/dfs/name</value>
<description>Path on the local filesystem where the NameNode stores the namespace and transactions logs persistently.</description>
</property>
<property>
<name>dfs.data.dir</name>
<value>/export/servers/hadoop/dfs/data</value>
<description>Comma separated list of paths on the localfilesystem of a DataNode where it should store its blocks.</description>
</property>

<!-- 指定HDFS副本的數量 -->
<property>
<name>dfs.replication</name>
<value>2</value>
</property>

<property>
<name>dfs.secondary.http.address</name>
<value>hadoop1:50090</value>
</property>
</configuration>

4、yarn-site.xml

<configuration>
<!-- 指定YARN的老大(ResourceManager)的地址 -->

<property>

<name>yarn.resourcemanager.hostname</name>
<value>hadoop1</value>
</property>
<!-- reducer獲取數據的方式 -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>

</configuration>

5、mapred-site.xml

<configuration>
<!-- 指定mr運行在yarn上 -->
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>

6、slaves

hadoop2
hadoop3

E、分發安裝目錄到其他機器節點

F、在namenode節點初始化HDFS 本例配置在 hadoop1 上

  執行 #./bin/hadoop namenode -format

G、啟動HDFS

  執行 # ./sbin/start-dfs.sh

[root@hadoop1 hadoop-2.8.4]# ./sbin/start-dfs.sh
Starting namenodes on [hadoop1]
hadoop1: namenode running as process 2343. Stop it first.
hadoop2: starting datanode, logging to /export/servers/hadoop-2.8.4/logs/hadoop-root-datanode-hadoop2.out
hadoop3: starting datanode, logging to /export/servers/hadoop-2.8.4/logs/hadoop-root-datanode-hadoop3.out
hadoop4: ssh: connect to host hadoop4 port 22: No route to host 【這個是我在slave 配置了hadoop4 ,然後我又沒有分發和啟動hadoop4節點,所以鏈接不到】
Starting secondary namenodes [hadoop1]
hadoop1: secondarynamenode running as process 2510. Stop it first. 【secondary namenode】hdfs 的冷備

H、啟動YARN

  執行 #./sbin/start-yarn.sh

starting yarn daemons
resourcemanager running as process 2697. Stop it first.【在哪臺機器上執行命令,resourcemanager就在這太機器上,然後再啟動slave配置的nodemanager】
hadoop2: starting nodemanager, logging to /export/servers/hadoop-2.8.4/logs/yarn-root-nodemanager-hadoop2.out
hadoop3: starting nodemanager, logging to /export/servers/hadoop-2.8.4/logs/yarn-root-nodemanager-hadoop3.out

二、測試

hadoop 集群安裝