1. 程式人生 > >0002-CENTOS7.2安裝CDH5.10和Kudu1.2

0002-CENTOS7.2安裝CDH5.10和Kudu1.2

Fayson的github: https://github.com/fayson/cdhproject

推薦關注微信公眾號:“Hadoop實操”,ID:gh_c4c535955d0f,或者掃描文末二維碼。

1.概述

本文件描述CENTOS7.2作業系統部署CDH企業版的過程。Cloudera企業級資料中心的安裝主要分為4個步驟:

1.叢集伺服器配置,包括安裝作業系統、關閉防火牆、同步伺服器時鐘等;
2.外部資料庫安裝
3.安裝Cloudera管理器;
4.安裝CDH叢集;
5.叢集完整性檢查,包括HDFS檔案系統、MapReduce、Hive等是否可以正常執行。

這篇文件將著重介紹Cloudera管理器與CDH的安裝,並基於以下假設:

1.作業系統版本:CENTOS7.2
2.MariaDB資料庫版本為10.2.1
3.CM版本:CDH 5.10.0
4.CDH版本:CDH 5.10.0
5.採用ec2-user對叢集進行部署
6.您已經下載CDH和CM的安裝包

2.前期準備

2.1.hostname及hosts配置

叢集中各個節點之間能互相通訊使用靜態IP地址。IP地址和主機名通過/etc/hosts配置,主機名/etc/hostname進行配置。

以cm節點(172.31.2.159)為例:

  • hostname配置

/etc/hostname檔案如下:

ip-172-31-2-159 

或者你可以通過命令修改立即生效

[email protected]
~$ sudo hostnamectl set-hostname ip-172-31-2-159

<font face="微軟雅黑" size=4 color=red >注意:這裡修改hostname跟REDHAT6的區別

  • hosts配置

/etc/hosts檔案如下:

172.31.2.159 ip-172-31-2-159
172.31.12.108 ip-172-31-12-108
172.31.5.236 ip-172-31-5-236
172.31.7.96 ip-172-31-7-96

以上兩步操作,在叢集中其它節點做相應配置。

2.2.禁用SELinux

在所有節點執行sudo setenforce 0 命令,此處使用批處理shell執行:

[email protected] ~$ sh ssh\_do\_all.sh  node.list "sudo setenforce 0" 

叢集所有節點修改/etc/selinux/config檔案如下:

SELINUX=disabled
SELINUXTYPE=targeted

2.3.關閉防火牆

叢集所有節點執行 sudo systemctl stop命令,此處通過shell批量執行命令如下:

[[email protected] ~]$ sh ssh_do_all.sh node.list "sudo systemctl stop firewalld"
[[email protected] ~]$ sh ssh_do_all.sh node.list "sudo systemctl disable firewalld"
[[email protected] ~]$ sh ssh_do_all.sh node.list "sudo systemctl status firewalld"

在這裡插入圖片描述

2.4.叢集時鐘同步

在CentOS7.2的作業系統上,已經預設的安裝了chrony,配置chrony時鐘同步,將cm(172.31.2.159)服務作為本地chrony伺服器,其它3臺伺服器與其保持同步,配置片段:

  • 172.31.2.159配置與自己同步

    [[email protected] ip-172-31-2-159 ~]$ sudo vim /etc/chrony.conf
    server ip-172-31-2-159 iburst
    #keyfile=/etc/chrony.keys
  • 叢集其它節點:在註釋下增加如下配置
[[email protected] ~]$ sudo vim /etc/chrony.conf
# Use public servers from the pool.ntp.org project.
# Please consider joining the pool (http://www.pool.ntp.org/join.html).
server ip-172-31-2-159 iburst
#keyfile=/etc/chrony.keys
  • 重啟所有機器的chrony服務
[[email protected] ~]$ sh ssh_do_all.sh node.list "sudo systemctl restart chronyd"

在這裡插入圖片描述

  • 驗證始終同步,在所有節點執行chronycsources命令,如下使用指令碼批量執行
[[email protected] ~]$ sh ssh_do_all.sh node.list "chronyc sources"

在這裡插入圖片描述

2.5.配置作業系統repo

  • 掛載作業系統iso檔案

    [[email protected] ~]$ sudo mkdir /media/DVD1
    [[email protected] ~]$ sudo mount -o loop 
    CentOS-7-x86_64-DVD-1611.iso /media/DVD1/

    在這裡插入圖片描述

  • 配置作業系統repo
    [[email protected] ~]$ sudo vim /etc/yum.repos.d/local_os.repo
    [local_iso] 
    name=CentOS-$releasever - Media
    baseurl=file:///media/DVD1
    gpgcheck=0
    enabled=1
    [[email protected] ~]$ sudo yum repolist

2.6.安裝http服務

  • 安裝httpd服務
[[email protected] ~]$ sudo yum -y install httpd
  • 啟動或停止httpd服務
[[email protected] ~]$ sudo systemctl start httpd
[[email protected] ~]$ sudo systemctl stop httpd
  • 安裝完httpd後,重新制作作業系統repo,換成http的方式方便其它伺服器也可以訪問
    [[email protected] ~]$ sudo mkdir /var/www/html/iso
    [[email protected] ~]$ sudo scp -r /media/DVD1/* /var/www/html/iso/
    [[email protected] ~]$ sudo vim /etc/yum.repos.d/os.repo 
    [osrepo]
    name=os_repo
    baseurl=http://172.31.2.159/iso/
    enabled=true
    gpgcheck=false
    [[email protected] ~]$ sudo yum repolist

2.7.安裝MariaDB

  • 由於centos7預設使用的是5.5.52版本的MariaDB,此處使用的10.2.1版本,點選,在官網下載rpm安裝包:
    MariaDB-10.2.1-centos7-x86_64-client.rpm
    MariaDB-10.2.1-centos7-x86_64-common.rpm
    MariaDB-10.2.1-centos7-x86_64-compat.rpm
    MariaDB-10.2.1-centos7-x86_64-server.rpm

將包下載到本地,放在同一目錄,執行createrepo命令生成rpm元資料。

此處使用apache2,將上述mariadb10.2.1目錄移動到/var/www/html目錄下, 使得使用者可以通過HTTP訪問這些rpm包。

[[email protected] ~]$ sudo mv mariadb10.2.1 /var/www/html/

安裝MariaDB依賴

[[email protected] ~]$ yum install libaio perl perl-DBI perl-Module-Pluggable perl-Pod-Escapes perl-Pod-Simple perl-libs perl-version

製作本地repo

[[email protected] ~]$ sudo vim /etc/yum.repos.d/mariadb.repo 
[mariadb]
name = MariaDB
baseurl = http://172.31.2.159/ mariadb10.2.1
enable = true
gpgcheck = false
[[email protected] ~]$ sudo yum repolist  
  • 安裝MariaDB

    [[email protected] ~]$ sudo yum -y install MariaDB-server MariaDB-client
  • 啟動並配置MariaDB
[[email protected] ~]$ sudo systemctl  start mariadb

[[email protected] ~]$ sudo /usr/bin/mysql_secure_installation

NOTE: RUNNING ALL PARTS OF THIS SCRIPT IS  RECOMMENDED FOR ALL MariaDB

       SERVERS IN PRODUCTION USE!   PLEASE READ EACH STEP CAREFULLY!

In order to log into MariaDB to secure it, we'll  need the current

password for the root user.  If you've just installed MariaDB, and

you haven't set the root password yet, the  password will be blank,

so you should just press enter here.

Enter current password for root (enter for none):  

OK, successfully used password, moving on...

Setting the root password ensures that nobody can  log into the MariaDB

root user without the proper authorisation.

Set root password? [Y/n] Y

New password:

Re-enter new password:

Password updated successfully!

Reloading privilege tables..

 ...  Success!

By default, a MariaDB installation has an  anonymous user, allowing anyone

to log into MariaDB without having to have a user  account created fo

them.  This  is intended only for testing, and to make the installation

go a bit smoother.  You should remove them before moving into a

production environment.

Remove anonymous users? [Y/n] Y

 ...  Success!

Normally, root should only be allowed to connect  from 'localhost'.  This

ensures that someone cannot guess at the root  password from the network.

Disallow root login remotely? [Y/n] n

 ...  skipping.

By default, MariaDB comes with a database named  'test' that anyone can

access.   This is also intended only for testing, and should be removed

before moving into a production environment.

Remove test database and access to it? [Y/n] Y

 - Dropping  test database...

 ...  Success!

 - Removing  privileges on test database...

 ...  Success!

Reloading the privilege tables will ensure that  all changes made so fa

will take effect immediately.

Reload privilege tables now? [Y/n] Y

 ...  Success!

Cleaning up...

All done!   If you've completed all of the above steps, your MariaDB

installation should now be secure.

Thanks for using MariaDB!
  • 建立CM和Hive需要的表
[[email protected] ~]$ mysql -uroot -p
Enter password: 
Welcome to the MariaDB monitor.  Commands end with ; or \g.
Your MariaDB connection id is 9
Server version: 10.2.1-MariaDB MariaDB Server
Copyright (c) 2000, 2016, Oracle, MariaDB Corporation Ab and others.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
MariaDB [(none)]>

create database metastore default character set utf8;
CREATE USER 'hive'@'%' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON metastore. * TO 'hive'@'%';
FLUSH PRIVILEGES;
create database cm default character set utf8;
CREATE USER 'cm'@'%' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON cm. * TO 'cm'@'%';
FLUSH PRIVILEGES;

create database am default character set utf8;
CREATE USER 'am'@'%' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON am. * TO 'am'@'%';
FLUSH PRIVILEGES;

create database rm default character set utf8;
CREATE USER 'rm'@'%' IDENTIFIED BY 'password';
GRANT ALL PRIVILEGES ON rm. * TO 'rm'@'%';
FLUSH PRIVILEGES;
  • 安裝jdbc驅動
[[email protected] ~]$ sudo mkdir -p /usr/share/java/
[[email protected] ~]$ sudo mv mysql-connector-java-5.1.37.jar /usr/share/java/
[[email protected] java]$ cd /usr/share/java
[[email protected] java]$ sudo ln -s mysql-connector-java-5.1.37.jar mysql-connector-java.jar
[[email protected] java]$ ll
total 964
-rw-r--r--. 1 root root 985600 Oct  6  2015 mysql-connector-java-5.1.37.jar
lrwxrwxrwx. 1 root root     31 Mar 29 14:37 mysql-connector-java.jar -> mysql-connector-java-5.1.37.jar

3.Cloudera Manager安裝

3.1.配置本地repo源

將Cloudera Manager安裝需要的7個rpm包下載到本地,放在同一目錄,執行createrepo命令生成rpm元資料。

[[email protected] cm]$ ls
cloudera-manager-agent-5.10.0-1.cm5100.p0.85.el7.x86_64.rpm
cloudera-manager-daemons-5.10.0-1.cm5100.p0.85.el7.x86_64.rpm
cloudera-manager-server-5.10.0-1.cm5100.p0.85.el7.x86_64.rpm
cloudera-manager-server-db-2-5.10.0-1.cm5100.p0.85.el7.x86_64.rpm
enterprise-debuginfo-5.10.0-1.cm5100.p0.85.el7.x86_64.rpm
jdk-6u31-linux-amd64.rpm
oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
[[email protected] cm]$ sudo createrepo .
Spawning worker 0 with 1 pkgs
Spawning worker 1 with 1 pkgs
Spawning worker 2 with 1 pkgs
Spawning worker 3 with 1 pkgs
Spawning worker 4 with 1 pkgs
Spawning worker 5 with 1 pkgs
Spawning worker 6 with 1 pkgs
Spawning worker 7 with 0 pkgs
Workers Finished
Saving Primary metadata
Saving file lists metadata
Saving other metadata
Generating sqlite DBs
Sqlite DBs complete
  • 配置Web伺服器

    此處使用apache2,將上述cdh5.10.0/cm5.10.0目錄移動到/var/www/html目錄下, 使得使用者可以通過HTTP訪問這些rpm包。

    [[email protected] ~]$ sudo mv cdh5.10.0/ cm5.10.0/ /var/www/html/

    在這裡插入圖片描述
    在這裡插入圖片描述

  • 製作Cloudera Manager的repo源
    [[email protected] ~]$ sudo vim /etc/yum.repos.d/cm.repo 
    [cmrepo]
    name = cm_repo
    baseurl = http://172.31.2.159/cm5.10.0.0
    enable = true
    gpgcheck = false
    [[email protected] yum.repos.d]$ sudo yum repolist
  • 驗證安裝JDK
    [[email protected] ~]$ sudo yum -y install oracle-j2sdk1.7-1.7.0+update67-1

3.2 安裝Cloudera Manager Server

  • 通過yum安裝ClouderaManager Server

    [[email protected] ~]$ sudo yum -y install cloudera-manager-server
  • 初始化資料庫

    [[email protected] ~]$ sudo /usr/share/cmf/schema/scm_prepare_database.sh mysql cm cm password
    JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera
    Verifying that we can write to /etc/cloudera-scm-server
    Creating SCM configuration file in /etc/cloudera-scm-server
    Executing:  /usr/java/jdk1.7.0_67-cloudera/bin/java -cp /usr/share/java/mysql-connector-java.jar:/usr/share/java/
    oracle-connector-java.jar:/usr/share/cmf/schema/../lib/* com.cloudera.enterprise.dbutil.DbCommandExecutor /etc/cloudera-scm-server/db.properties com.cloudera.cmf.db.
    [                          main] DbCommandExecutor              INFO  Successfully connected to database.
    All done, your SCM database is configured correctly!
  • 啟動Cloudera Manager Server

    [[email protected] ~]$ sudo systemctl start cloudera-scm-server
  • 檢查埠是否監聽

    [[email protected] ~]$ sudo netstat -lnpt | grep 7180
    tcp        0      0 0.0.0.0:7180            0.0.0.0:*               LISTEN      6890/java  
  • 通過http://172.31.2.159:7180/cmf/login訪問CM
    在這裡插入圖片描述

4.CDH安裝

4.1.CDH叢集安裝嚮導

1.admin/admin登入到CM
2.同意license協議,點選繼續
在這裡插入圖片描述

3.選擇60試用,點選繼續
在這裡插入圖片描述

4.點選“繼續”
在這裡插入圖片描述

5.輸入主機ip或者名稱,點選搜尋找到主機後點擊繼續
在這裡插入圖片描述

在這裡插入圖片描述

6.點選“繼續”
在這裡插入圖片描述

7.使用parcel選擇,點選“更多選項”,點選“-”刪除其它所有地址,輸入http://172.31.2.159/cm5.10.0/,點選“儲存更改”

在這裡插入圖片描述

8.選擇自定義儲存庫,輸入cm的http地址

在這裡插入圖片描述

9.點選“繼續”,進入下一步安裝jdk

在這裡插入圖片描述

10.點選“繼續”,進入下一步,預設多使用者模式

在這裡插入圖片描述

11.點選“繼續”,進入下一步配置ssh賬號密碼

在這裡插入圖片描述

12.點選“繼續”,進入下一步,安裝Cloudera Manager相關到各個節點

在這裡插入圖片描述

在這裡插入圖片描述

13.點選“繼續”,進入下一步安裝cdh到各個節點

在這裡插入圖片描述

在這裡插入圖片描述

14.點選“繼續”,進入下一步主機檢查,確保所有檢查項均通過

在這裡插入圖片描述
在這裡插入圖片描述

點選完成進入服務安裝嚮導。

4.2.叢集設定安裝嚮導

1.選擇需要安裝的服務

在這裡插入圖片描述

2.點選“繼續”,進入叢集角色分配

在這裡插入圖片描述

3.點選“繼續”,進入下一步,測試資料庫連線

在這裡插入圖片描述

4.測試成功,點選“繼續”,進入目錄設定,此處使用預設預設目錄,根據實際情況進行目錄修改

在這裡插入圖片描述

5.點選“繼續”,進入各個服務啟動

在這裡插入圖片描述

6.安裝成功

在這裡插入圖片描述

7.安裝成功後進入home管理介面

在這裡插入圖片描述

5.Kudu安裝

CDH從5.10開始,打包整合Kudu1.2,並且Cloudera正式提供支援。這個版本開始Kudu的安裝較之前要簡單很多,省去了Impala_Kudu,安裝完Kudu,Impala即可直接操作Kudu。

以下安裝步驟基於使用者使用Cloudera Manager來安裝和部署Kudu1.2

5.1.安裝csd檔案

1.下載csd檔案

[[email protected] ~]# wget http://archive.cloudera.com/kudu/csd/KUDU-5.10.0.jar

2.將下載的jar包檔案移動到/opt/cloudera/csd目錄

[[email protected] ~]# mv KUDU-5.10.0.jar /opt/cloudera/csd

3.修改許可權

[[email protected] ~]# chown cloudera-scm:cloudera-scm /opt/cloudera/csd/KUDU-5.10.0.jar 
[[email protected] ~]# chmod 644 /opt/cloudera/csd/KUDU-5.10.0.jar

4.重啟Cloudera Manager服務

[[email protected] ~]# systemctl restart cloudera-scm-server

5.2.安裝Kudu服務

1.下載Kudu服務需要的Parcel包

[[email protected] ~]# wget http://archive.cloudera.com/kudu/parcels/5.10/KUDU-1.2.0-1.cdh5.10.1.p0.66-el7.parcel
[[email protected] ~]# wget http://archive.cloudera.com/kudu/parcels/5.10/KUDU-1.2.0-1.cdh5.10.1.p0.66-el7.parcel.sha1
[[email protected] ~]# wget http://archive.cloudera.com/kudu/parcels/5.10/manifest.json

2.將Kudu的Parcel包部署到http服務

[[email protected] ~]# mkdir kudu1.2
[[email protected] ~]# mv KUDU-1.2.0-1.cdh5.10.1.p0.66-el7.parcel* kudu1.2/
[[email protected] ~]# mv manifest.json kudu1.2
[[email protected] ~]# mv kudu1.2/ /var/www/html/
[[email protected] ~]# systemctl start httpd

3.檢查http顯示Kudu正常:

在這裡插入圖片描述

4.通過CM介面配置Kudu的Parcel地址,並下載,分發,啟用Kudu。
在這裡插入圖片描述
在這裡插入圖片描述

5.通過CM安裝Kudu1.2

新增Kudu服務

在這裡插入圖片描述

選擇Master和Tablet Server

在這裡插入圖片描述

配置相應的目錄,<font face="微軟雅黑" size=4 color=red >注:無論是Master還是Tablet根據實際情況資料目錄(fs_data_dir)應該都可能有多個,以提高併發讀寫,從而提高Kudu效能
在這裡插入圖片描述

啟動Kudu服務

0002-CENTOS7.2安裝CDH5.10和Kudu1.2

安裝完畢

在這裡插入圖片描述

5.3.配置Impala

在CDH5.10中,安裝完Kudu1.2後,預設Impala即可直接操作Kudu進行SQL操作,但為了省去每次建表都需要在TBLPROPERTIES中新增kudu_master_addresses屬性,建議在Impala的高階配置Kudu Master的地址:

--kudu\_master\_hosts=ip-172-31-2-159:7051

在這裡插入圖片描述

6.快速元件服務驗證

6.1.HDFS驗證(mkdir+put+cat+get)

[[email protected] ~]# hadoop fs -mkdir -p /lilei/test_table
[[email protected] ~]# cat > a.txt
1#2
c#d
我#你^C
[[email protected] ~]# 
[[email protected] ~]# 
[[email protected] ~]# 
[[email protected] ~]# hadoop fs -put a.txt /lilei/test_table
[[email protected] ~]# hadoop fs -cat /lilei/test_table/a.txt
1#2
c#d
[[email protected] ~]# rm -rf a.txt
[[email protected] ~]# 
[[email protected] ~]# hadoop fs -get /lilei/test_table/a.txt
[[email protected] ~]# 
[[email protected] ~]# cat a.txt
1#2
c#d

6.2.Hive驗證

[[email protected] ~]# hive

Logging initialized using configuration in jar:file:/opt/cloudera/parcels/CDH-5.10.0-1.cdh5.10.0.p0.41/jars/hive-common-1.1.0-cdh5.10.0.jar!/hive-log4j.properties
WARNING: Hive CLI is deprecated and migration to Beeline is recommended.
hive> create external table test_table
    > (
    > s1 string,
    > s2 string
    > )
    > row format delimited fields terminated by '#'
    > stored as textfile location '/lilei/test_table';
OK
Time taken: 0.631 seconds
hive> select * from test_table;
OK
1   2
c   d
Time taken: 0.36 seconds, Fetched: 2 row(s)
hive> select count(*) from test_table;
Query ID = root_20170404013939_69844998-4456-4bc1-9da5-53ea91342e43
Total jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>
Starting Job = job_1491283979906_0005, Tracking URL = http://ip-172-31-2-159:8088/proxy/application_1491283979906_0005/
Kill Command = /opt/cloudera/parcels/CDH-5.10.0-1.cdh5.10.0.p0.41/lib/hadoop/bin/hadoop job  -kill job_1491283979906_0005
Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 1
2017-04-04 01:39:25,425 Stage-1 map = 0%,  reduce = 0%
2017-04-04 01:39:31,689 Stage-1 map = 100%,  reduce = 0%, Cumulative CPU 1.02 sec
2017-04-04 01:39:36,851 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 2.34 sec
MapReduce Total cumulative CPU time: 2 seconds 340 msec
Ended Job = job_1491283979906_0005
MapReduce Jobs Launched: 
Stage-Stage-1: Map: 1  Reduce: 1   Cumulative CPU: 2.34 sec   HDFS Read: 6501 HDFS Write: 2 SUCCESS
Total MapReduce CPU Time Spent: 2 seconds 340 msec
OK
2
Time taken: 21.56 seconds, Fetched: 1 row(s)

6.3.MapReduce驗證

[[email protected] ~]# hadoop jar /opt/cloudera/parcels/CDH/lib/hadoop-0.20-mapreduce/hadoop-examples.jar pi 5 5
Number of Maps  = 5
Samples per Map = 5
Wrote input for Map #0
Wrote input for Map #1
Wrote input for Map #2
Wrote input for Map #3
Wrote input for Map #4
Starting Job
17/04/04 01:38:15 INFO client.RMProxy: Connecting to ResourceManager at ip-172-31-2-159/172.31.2.159:8032
17/04/04 01:38:15 INFO mapreduce.JobSubmissionFiles: Permissions on staging directory /user/root/.staging are incorrect: rwxrwxrwx. Fixing permissions to correct value rwx------
17/04/04 01:38:15 INFO input.FileInputFormat: Total input paths to process : 5
17/04/04 01:38:15 INFO mapreduce.JobSubmitter: number of splits:5
17/04/04 01:38:15 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1491283979906_0004
17/04/04 01:38:16 INFO impl.YarnClientImpl: Submitted application application_1491283979906_0004
17/04/04 01:38:16 INFO mapreduce.Job: The url to track the job: http://ip-172-31-2-159:8088/proxy/application_1491283979906_0004/
17/04/04 01:38:16 INFO mapreduce.Job: Running job: job_1491283979906_0004
17/04/04 01:38:21 INFO mapreduce.Job: Job job_1491283979906_0004 running in uber mode : false
17/04/04 01:38:21 INFO mapreduce.Job:  map 0% reduce 0%
17/04/04 01:38:26 INFO mapreduce.Job:  map 100% reduce 0%
17/04/04 01:38:32 INFO mapreduce.Job:  map 100% reduce 100%
17/04/04 01:38:32 INFO mapreduce.Job: Job job_1491283979906_0004 completed successfully
17/04/04 01:38:32 INFO mapreduce.Job: Counters: 49
    File System Counters
        FILE: Number of bytes read=64
        FILE: Number of bytes written=749758
        FILE: Number of read operations=0
        FILE: Number of large read operations=0
        FILE: Number of write operations=0
        HDFS: Number of bytes read=1350
        HDFS: Number of bytes written=215
        HDFS: Number of read operations=23
        HDFS: Number of large read operations=0
        HDFS: Number of write operations=3
    Job Counters 
        Launched map tasks=5
        Launched reduce tasks=1
        Data-local map tasks=5
        Total time spent by all maps in occupied slots (ms)=16111
        Total time spent by all reduces in occupied slots (ms)=2872
        Total time spent by all map tasks (ms)=16111
        Total time spent by all reduce tasks (ms)=2872
        Total vcore-seconds taken by all map tasks=16111
        Total vcore-seconds taken by all reduce tasks=2872
        Total megabyte-seconds taken by all map tasks=16497664
        Total megabyte-seconds taken by all reduce tasks=2940928
    Map-Reduce Framework
        Map input records=5
        Map output records=10
        Map output bytes=90
        Map output materialized bytes=167
        Input split bytes=760
        Combine input records=0
        Combine output records=0
        Reduce input groups=2
        Reduce shuffle bytes=167
        Reduce input records=10
        Reduce output records=0
        Spilled Records=20
        Shuffled Maps =5
        Failed Shuffles=0
        Merged Map outputs=5
        GC time elapsed (ms)=213
        CPU time spent (ms)=3320
        Physical memory (bytes) snapshot=2817884160
        Virtual memory (bytes) snapshot=9621606400
        Total committed heap usage (bytes)=2991587328
    Shuffle Errors
        BAD_ID=0
        CONNECTION=0
        IO_ERROR=0
        WRONG_LENGTH=0
        WRONG_MAP=0
        WRONG_REDUCE=0
    File Input Format Counters 
        Bytes Read=590
    File Output Format Counters 
        Bytes Written=97
Job Finished in 17.145 seconds
Estimated value of Pi is 3.68000000000000000000

6.4.Impala驗證

[[email protected] ~]# impala-shell -i ip-172-31-7-96
Starting Impala Shell without Kerberos authentication
Connected to ip-172-31-7-96:21000
Server version: impalad version 2.7.0-cdh5.10.0 RELEASE (build 785a073cd07e2540d521ecebb8b38161ccbd2aa2)
***********************************************************************************
Welcome to the Impala shell.
(Impala Shell v2.7.0-cdh5.10.0 (785a073) built on Fri Jan 20 12:03:56 PST 2017)

Run the PROFILE command after a query has finished to see a comprehensive summary
of all the performance and diagnostic information that Impala gathered for that
query. Be warned, it can be very long!
***********************************************************************************
[ip-172-31-7-96:21000] > show tables;
Query: show tables
+------------+
| name       |
+------------+
| test_table |
+------------+
Fetched 1 row(s) in 0.20s
[ip-172-31-7-96:21000] > select * from test_table;
Query: select * from test_table
Query submitted at: 2017-04-04 01:41:56 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=c4a06bd46f9106b:4a69f04800000000
+----+----+
| s1 | s2 |
+----+----+
| 1  | 2  |
| c  | d  |
+----+----+
Fetched 2 row(s) in 3.73s
[ip-172-31-7-96:21000] > select count(*) from test_table;
Query: select count(*) from test_table
Query submitted at: 2017-04-04 01:42:06 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=2a415724696f7414:1f9113ea00000000
+----------+
| count(*) |
+----------+
| 2        |
+----------+
Fetched 1 row(s) in 0.15s

6.5.Spark驗證

[[email protected] ~]# spark-shell
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel).
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 1.6.0
      /_/

Using Scala version 2.10.5 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_67)
Type in expressions to have them evaluated.
Type :help for more information.
Spark context available as sc (master = yarn-client, app id = application_1491283979906_0006).
17/04/04 01:43:26 WARN metastore.ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.1.0
17/04/04 01:43:27 WARN metastore.ObjectStore: Failed to get database default, returning NoSuchObjectException
SQL context available as sqlContext.

scala> var textFile=sc.textFile("hdfs://ip-172-31-2-159:8020/lilei/test_table/a.txt")
textFile: org.apache.spark.rdd.RDD[String] = hdfs://ip-172-31-2-159:8020/lilei/test_table/a.txt MapPartitionsRDD[1] at textFile at <console>:27

scala> 

scala> textFile.count()
res0: Long = 2

6.6.Kudu驗證

[[email protected] ~]# impala-shell -i ip-172-31-7-96
Starting Impala Shell without Kerberos authentication
Connected to ip-172-31-7-96:21000
Server version: impalad version 2.7.0-cdh5.10.0 RELEASE (build 785a073cd07e2540d521ecebb8b38161ccbd2aa2)
***********************************************************************************
Welcome to the Impala shell.
(Impala Shell v2.7.0-cdh5.10.0 (785a073) built on Fri Jan 20 12:03:56 PST 2017)

Every command must be terminated by a ';'.
***********************************************************************************
[ip-172-31-7-96:21000] > CREATE TABLE my_first_table
                       > (
                       >   id BIGINT,
                       >   name STRING,
                       >   PRIMARY KEY(id)
                       > )
                       > PARTITION BY HASH PARTITIONS 16
                       > STORED AS KUDU;
Query: create TABLE my_first_table
(
  id BIGINT,
  name STRING,
  PRIMARY KEY(id)
)
PARTITION BY HASH PARTITIONS 16
STORED AS KUDU

Fetched 0 row(s) in 1.35s
[ip-172-31-7-96:21000] > INSERT INTO my_first_table VALUES (99, "sarah");
Query: insert INTO my_first_table VALUES (99, "sarah")
Query submitted at: 2017-04-04 01:46:08 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=824ce0b3765c6b91:5ea8dd7c00000000
Modified 1 row(s), 0 row error(s) in 3.37s
[ip-172-31-7-96:21000] > 
[ip-172-31-7-96:21000] > INSERT INTO my_first_table VALUES (1, "john"), (2, "jane"), (3, "jim");
Query: insert INTO my_first_table VALUES (1, "john"), (2, "jane"), (3, "jim")
Query submitted at: 2017-04-04 01:46:13 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=a645259c3b8ae7cd:e446e15500000000
Modified 3 row(s), 0 row error(s) in 0.11s
[ip-172-31-7-96:21000] > select * from my_first_table;
Query: select * from my_first_table
Query submitted at: 2017-04-04 01:46:19 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=f44021589ff0d94d:8d30568200000000
+----+-------+
| id | name  |
+----+-------+
| 2  | jane  |
| 3  | jim   |
| 1  | john  |
| 99 | sarah |
+----+-------+
Fetched 4 row(s) in 0.55s
[ip-172-31-7-96:21000] > delete from my_first_table where id =99;
Query: delete from my_first_table where id =99
Query submitted at: 2017-04-04 01:46:56 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=814090b100fdf0b4:1b516fe400000000
Modified 1 row(s), 0 row error(s) in 0.15s
[ip-172-31-7-96:21000] > 
[ip-172-31-7-96:21000] > select * from my_first_table;
Query: select * from my_first_table
Query submitted at: 2017-04-04 01:46:57 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=724aa3f84cedb109:a679bf0200000000
+----+------+
| id | name |
+----+------+
| 2  | jane |
| 3  | jim  |
| 1  | john |
+----+------+
Fetched 3 row(s) in 0.15s
[ip-172-31-7-96:21000] > INSERT INTO my_first_table VALUES (99, "sarah");
Query: insert INTO my_first_table VALUES (99, "sarah")
Query submitted at: 2017-04-04 01:47:32 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=6244b3c6d33b443e:f43c857300000000
Modified 1 row(s), 0 row error(s) in 0.11s
[ip-172-31-7-96:21000] > 
[ip-172-31-7-96:21000] > update my_first_table set name='lilei' where id=99;
Query: update my_first_table set name='lilei' where id=99
Query submitted at: 2017-04-04 01:47:32 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=8f4ab0dd3c19f9df:b2c7bdfa00000000
Modified 1 row(s), 0 row error(s) in 0.13s
[ip-172-31-7-96:21000] > select * from my_first_table;
Query: select * from my_first_table
Query submitted at: 2017-04-04 01:47:34 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=6542579c8bd5b6ad:af68f50800000000
+----+-------+
| id | name  |
+----+-------+
| 2  | jane  |
| 3  | jim   |
| 1  | john  |
| 99 | lilei |
+----+-------+
Fetched 4 row(s) in 0.15s
[ip-172-31-7-96:21000] > upsert  into my_first_table values(1, "john"), (4, "tom"), (99, "lilei1");
Query: upsert into my_first_table values(1, "john"), (4, "tom"), (99, "lilei1")
Query submitted at: 2017-04-04 01:48:52 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=694fc7ac2bc71d21:947f1fa200000000
Modified 3 row(s), 0 row error(s) in 0.11s
[ip-172-31-7-96:21000] > 
[ip-172-31-7-96:21000] > select * from my_first_table;
Query: select * from my_first_table
Query submitted at: 2017-04-04 01:48:52 (Coordinator: http://ip-172-31-7-96:25000)
Query progress can be monitored at: http://ip-172-31-7-96:25000/query_plan?query_id=a64e0ee707762b6b:69248a6c00000000
+----+--------+
| id | name   |
+----+--------+
| 2  | jane   |
| 3  | jim    |
| 1  | john   |
| 99 | lilei1 |
| 4  | tom    |
+----+--------+
Fetched 5 row(s) in 0.16s

為天地立心,為生民立命,為往聖繼絕學,為萬世開太平。

推薦關注Hadoop實操,第一時間,分享更多Hadoop乾貨,歡迎轉發和分享。
在這裡插入圖片描述

原創文章,歡迎轉載,轉載請註明:轉載自微信公眾號Hadoop實操

0002-CENTOS7.2安裝CDH5.10和Kudu1.2