1. 程式人生 > >HBASE-1.1.12叢集啟動之後Hmaster,HregionServer程序會很快死掉問題

HBASE-1.1.12叢集啟動之後Hmaster,HregionServer程序會很快死掉問題

一、檢視日誌資訊如下

  2017-11-25 10:37:25,646 INFO  [main] zookeeper.ZooKeeper: Client environment:java.library.path=/usr/local/hadoop-2.7.4/lib/native
2017-11-25 10:37:25,646 INFO  [main] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp
2017-11-25 10:37:25,646 INFO  [main] zookeeper.ZooKeeper: Client environment:java.compiler=<NA>
2017-11-25 10:37:25,647 INFO  [main] zookeeper.ZooKeeper: Client environment:os.name=Linux
2017-11-25 10:37:25,647 INFO  [main] zookeeper.ZooKeeper: Client environment:os.arch=amd64
2017-11-25 10:37:25,647 INFO  [main] zookeeper.ZooKeeper: Client environment:os.version=3.10.0-514.el7.x86_64
2017-11-25 10:37:25,647 INFO  [main] zookeeper.ZooKeeper: Client environment:user.name=root
2017-11-25 10:37:25,647 INFO  [main] zookeeper.ZooKeeper: Client environment:user.home=/root
2017-11-25 10:37:25,648 INFO  [main] zookeeper.ZooKeeper: Client environment:user.dir=/root
2017-11-25 10:37:25,650 INFO  [main] zookeeper.ZooKeeper: Initiating client connection, connectString=master:2181,slave1:2181,slave2:2181 sessionTimeout=90000 watcher=master:600000x0, quorum=master:2181,slave1:2181,slave2:2181, baseZNode=/hbase
2017-11-25 10:37:25,859 INFO  [main-SendThread(slave2:2181)] zookeeper.ClientCnxn: Opening socket connection to server slave2/172.16.169.12:2181. Will not attempt to authenticate using SASL (unknown error)
2017-11-25 10:37:25,927 INFO  [main-SendThread(slave2:2181)] zookeeper.ClientCnxn: Socket connection established to slave2/172.16.169.12:2181, initiating session
2017-11-25 10:37:26,096 INFO  [main-SendThread(slave2:2181)] zookeeper.ClientCnxn: Unable to read additional data from server sessionid 0x0, likely server has closed socket, closing socket connection and attempting reconnect
2017-11-25 10:37:26,240 WARN  [main] zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper, quorum=master:2181,slave1:2181,slave2:2181, exception=org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /hbase
2017-11-25 10:37:26,871 INFO  [main-SendThread(master:2181)] zookeeper.ClientCnxn: Opening socket connection to server master/172.16.169.10:2181. Will not attempt to authenticate using SASL (unknown error)
2017-11-25 10:37:26,873 INFO  [main-SendThread(master:2181)] zookeeper.ClientCnxn: Socket connection established to master/172.16.169.10:2181, initiating session
2017-11-25 10:37:26,905 INFO  [main-SendThread(master:2181)] zookeeper.ClientCnxn: Session establishment complete on server master/172.16.169.10:2181, sessionid = 0x15ff10795050002, negotiated timeout = 40000
2017-11-25 10:37:27,708 INFO  [RpcServer.responder] ipc.RpcServer: RpcServer.responder: starting
2017-11-25 10:37:27,731 INFO  [RpcServer.listener,port=60000] ipc.RpcServer: RpcServer.listener,port=60000: starting
2017-11-25 10:37:29,361 INFO  [main] mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2017-11-25 10:37:29,412 INFO  [main] http.HttpRequestLog: Http request log for http.requests.master is not defined
2017-11-25 10:37:29,585 INFO  [main] http.HttpServer: Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter)
2017-11-25 10:37:29,636 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context master
2017-11-25 10:37:29,638 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2017-11-25 10:37:29,639 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2017-11-25 10:37:29,860 INFO  [main] http.HttpServer: Jetty bound to port 60010
2017-11-25 10:37:29,860 INFO  [main] mortbay.log: jetty-6.1.26
2017-11-25 10:37:34,181 INFO  [main] mortbay.log: Started

[email protected]:60010
2017-11-25 10:37:34,210 INFO  [main] master.HMaster: hbase.rootdir=hdfs://master:9000/hbase, hbase.cluster.distributed=true
2017-11-25 10:37:34,375 INFO  [main] master.HMaster: Adding backup master ZNode /hbase/backup-masters/master,60000,1511577430262
2017-11-25 10:37:34,997 INFO  [master:60000.activeMasterManager] master.ActiveMasterManager: Deleting ZNode for /hbase/backup-masters/master,60000,1511577430262 from backup master directory
2017-11-25 10:37:35,154 INFO  [master:60000.activeMasterManager] master.ActiveMasterManager: Registered Active Master=master,60000,1511577430262
2017-11-25 10:37:36,813 INFO  [master/master/172.16.169.10:60000] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x6ffbec93 connecting to ZooKeeper ensemble=master:2181,slave1:2181,slave2:2181
2017-11-25 10:37:36,835 INFO  [master/master/172.16.169.10:60000] zookeeper.ZooKeeper: Initiating client connection, connectString=master:2181,slave1:2181,slave2:2181 sessionTimeout=90000 watcher=hconnection-0x6ffbec930x0, quorum=master:2181,slave1:2181,slave2:2181, baseZNode=/hbase
2017-11-25 10:37:36,932 INFO  [master/master/172.16.169.10:60000-SendThread(master:2181)] zookeeper.ClientCnxn: Opening socket connection to server master/172.16.169.10:2181. Will not attempt to authenticate using SASL (unknown error)
2017-11-25 10:37:36,936 INFO  [master/master/172.16.169.10:60000-SendThread(master:2181)] zookeeper.ClientCnxn: Socket connection established to master/172.16.169.10:2181, initiating session
2017-11-25 10:37:37,017 INFO  [master/master/172.16.169.10:60000-SendThread(master:2181)] zookeeper.ClientCnxn: Session establishment complete on server master/172.16.169.10:2181, sessionid = 0x15ff10795050004, negotiated timeout = 40000
2017-11-25 10:37:37,431 INFO  [master/master/172.16.169.10:60000] regionserver.HRegionServer: ClusterId : 54fafe3b-fc82-40f5-9791-47ef66133d4f
2017-11-25 10:37:43,571 INFO  [master:60000.activeMasterManager] fs.HFileSystem: Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks
2017-11-25 10:37:43,649 INFO  [master:60000.activeMasterManager] coordination.SplitLogManagerCoordination: Found 0 orphan tasks and 0 rescan nodes
2017-11-25 10:37:43,932 INFO  [master:60000.activeMasterManager] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x157f9d2f connecting to ZooKeeper ensemble=master:2181,slave1:2181,slave2:2181
2017-11-25 10:37:43,933 INFO  [master:60000.activeMasterManager] zookeeper.ZooKeeper: Initiating client connection, connectString=master:2181,slave1:2181,slave2:2181 sessionTimeout=90000 watcher=hconnection-0x157f9d2f0x0, quorum=master:2181,slave1:2181,slave2:2181, baseZNode=/hbase
2017-11-25 10:37:43,949 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Opening socket connection to server slave1/172.16.169.11:2181. Will not attempt to authenticate using SASL (unknown error)
2017-11-25 10:37:43,956 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Socket connection established to slave1/172.16.169.11:2181, initiating session
2017-11-25 10:37:43,964 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Session establishment complete on server slave1/172.16.169.11:2181, sessionid = 0x25ff10795560003, negotiated timeout = 40000
2017-11-25 10:37:44,063 INFO  [master:60000.activeMasterManager] balancer.StochasticLoadBalancer: loading config
2017-11-25 10:37:44,365 INFO  [master:60000.activeMasterManager] master.HMaster: Server active/primary master=master,60000,1511577430262, sessionid=0x15ff10795050002, setting cluster-up flag (Was=true)
2017-11-25 10:37:44,609 INFO  [master:60000.activeMasterManager] procedure.ZKProcedureUtil: Clearing all procedure znodes: /hbase/flush-table-proc/acquired /hbase/flush-table-proc/reached /hbase/flush-table-proc/abort
2017-11-25 10:37:44,732 INFO  [master:60000.activeMasterManager] procedure.ZKProcedureUtil: Clearing all procedure znodes: /hbase/online-snapshot/acquired /hbase/online-snapshot/reached /hbase/online-snapshot/abort
2017-11-25 10:37:45,007 INFO  [master:60000.activeMasterManager] master.MasterCoprocessorHost: System coprocessor loading is enabled
2017-11-25 10:37:45,189 INFO  [master:60000.activeMasterManager] procedure2.ProcedureExecutor: Starting procedure executor threads=5
2017-11-25 10:37:45,191 INFO  [master:60000.activeMasterManager] wal.WALProcedureStore: Starting WAL Procedure Store lease recovery
2017-11-25 10:37:45,207 INFO  [master:60000.activeMasterManager] util.FSHDFSUtils: Recovering lease on dfs file hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000004.log
2017-11-25 10:37:45,308 INFO  [master:60000.activeMasterManager] util.FSHDFSUtils: recoverLease=false, attempt=0 on file=hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000004.log after 100ms
2017-11-25 10:37:49,317 INFO  [master:60000.activeMasterManager] util.FSHDFSUtils: recoverLease=true, attempt=1 on file=hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000004.log after 4109ms
2017-11-25 10:37:49,414 WARN  [master:60000.activeMasterManager] wal.WALProcedureStore: Unable to read tracker for hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000004.log - Invalid Trailer version. got 202 expected 1
2017-11-25 10:37:49,503 INFO  [master:60000.activeMasterManager] wal.WALProcedureStore: Lease acquired for flushLogId: 5
2017-11-25 10:37:49,646 WARN  [master:60000.activeMasterManager] wal.ProcedureWALFormatReader: nothing left to decode. exiting with missing EOF
2017-11-25 10:37:49,785 INFO  [master:60000.activeMasterManager] zookeeper.RecoverableZooKeeper: Process identifier=replicationLogCleaner connecting to ZooKeeper ensemble=master:2181,slave1:2181,slave2:2181
2017-11-25 10:37:49,785 INFO  [master:60000.activeMasterManager] zookeeper.ZooKeeper: Initiating client connection, connectString=master:2181,slave1:2181,slave2:2181 sessionTimeout=90000 watcher=replicationLogCleaner0x0, quorum=master:2181,slave1:2181,slave2:2181, baseZNode=/hbase
2017-11-25 10:37:49,829 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Opening socket connection to server slave1/172.16.169.11:2181. Will not attempt to authenticate using SASL (unknown error)
2017-11-25 10:37:49,841 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Socket connection established to slave1/172.16.169.11:2181, initiating session
2017-11-25 10:37:49,864 INFO  [master:60000.activeMasterManager-SendThread(slave1:2181)] zookeeper.ClientCnxn: Session establishment complete on server slave1/172.16.169.11:2181, sessionid = 0x25ff10795560004, negotiated timeout = 40000
2017-11-25 10:37:49,899 INFO  [master:60000.activeMasterManager] master.ServerManager: Waiting for region servers count to settle; currently checked in 0, slept for 0 ms, expecting minimum of 1, maximum of 2147483647, timeout of 4500 ms, interval of 1500 ms.
2017-11-25 10:37:51,441 INFO  [master:60000.activeMasterManager] master.ServerManager: Waiting for region servers count to settle; currently checked in 0, slept for 1542 ms, expecting minimum of 1, maximum of 2147483647, timeout of 4500 ms, interval of 1500 ms.
2017-11-25 10:37:52,378 INFO  [PriorityRpcServer.handler=5,queue=1,port=60000] master.ServerManager: Registering server=slave1,16020,1511577422976
2017-11-25 10:37:52,399 INFO  [PriorityRpcServer.handler=3,queue=1,port=60000] master.ServerManager: Registering server=slave2,16020,1511577423313
2017-11-25 10:37:52,399 INFO  [PriorityRpcServer.handler=1,queue=1,port=60000] master.ServerManager: Registering server=master,16020,1511577430553
2017-11-25 10:37:52,430 INFO  [master:60000.activeMasterManager] master.ServerManager: Waiting for region servers count to settle; currently checked in 3, slept for 2531 ms, expecting minimum of 1, maximum of 2147483647, timeout of 4500 ms, interval of 1500 ms.
2017-11-25 10:37:52,567 INFO  [WALProcedureStoreSyncThread] wal.WALProcedureStore: Remove log: hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000004.log
2017-11-25 10:37:52,568 INFO  [WALProcedureStoreSyncThread] wal.WALProcedureStore: Removed logs: [hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000005.log, hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000006.log]
2017-11-25 10:37:52,599 INFO  [WALProcedureStoreSyncThread] wal.WALProcedureStore: Remove log: hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000005.log
2017-11-25 10:37:52,600 INFO  [WALProcedureStoreSyncThread] wal.WALProcedureStore: Removed logs: [hdfs://master:9000/hbase/MasterProcWALs/state-00000000000000000006.log]
2017-11-25 10:37:53,933 INFO  [master:60000.activeMasterManager] master.ServerManager: Waiting for region servers count to settle; currently checked in 3, slept for 4034 ms, expecting minimum of 1, maximum of 2147483647, timeout of 4500 ms, interval of 1500 ms.
2017-11-25 10:37:54,415 INFO  [master:60000.activeMasterManager] master.ServerManager: Finished waiting for region servers count to settle; checked in 3, slept for 4516 ms, expecting minimum of 1, maximum of 2147483647, master is running
2017-11-25 10:37:54,470 INFO  [master:60000.activeMasterManager] master.MasterFileSystem: Log folder hdfs://master:9000/hbase/WALs/master,16020,1511532189674 doesn't belong to a known region server, splitting
2017-11-25 10:37:54,528 INFO  [master:60000.activeMasterManager] master.MasterFileSystem: Log folder hdfs://master:9000/hbase/WALs/master,16020,1511577430553 belongs to an existing region server
2017-11-25 10:37:54,581 INFO  [master:60000.activeMasterManager] master.MasterFileSystem: Log folder hdfs://master:9000/hbase/WALs/slave1,16020,1511532206980 doesn't belong to a known region server, splitting
2017-11-25 10:37:54,618 INFO  [master:60000.activeMasterManager] master.MasterFileSystem: Log folder hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830 doesn't belong to a known region server, splitting
2017-11-25 10:37:54,805 INFO  [master:60000.activeMasterManager] master.SplitLogManager: dead splitlog workers [slave2,16020,1511532219830]
2017-11-25 10:37:54,855 INFO  [master:60000.activeMasterManager] master.SplitLogManager: started splitting 1 logs in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] for [slave2,16020,1511532219830]
2017-11-25 10:37:55,654 INFO  [master,60000,1511577430262_splitLogManager__ChoreService_1] master.SplitLogManager: total tasks = 1 unassigned = 1 tasks={/hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830..meta.1511532416907.meta=last_update = -1 last_version = -1 cur_worker_name = null status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0}
2017-11-25 10:38:00,482 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830..meta.1511532416907.meta acquired by master,16020,1511577430553
2017-11-25 10:38:00,658 INFO  [master,60000,1511577430262_splitLogManager__ChoreService_1] master.SplitLogManager: total tasks = 1 unassigned = 0 tasks={/hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830..meta.1511532416907.meta=last_update = 1511577480483 last_version = 1 cur_worker_name = master,16020,1511577430553 status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0}
2017-11-25 10:38:06,250 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830..meta.1511532416907.meta entered state: DONE master,16020,1511577430553
2017-11-25 10:38:06,464 INFO  [main-EventThread] wal.WALSplitter: Archived processed log hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting/slave2%2C16020%2C1511532219830..meta.1511532416907.meta to hdfs://master:9000/hbase/oldWALs/slave2%2C16020%2C1511532219830..meta.1511532416907.meta
2017-11-25 10:38:06,471 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: Done splitting /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830..meta.1511532416907.meta
2017-11-25 10:38:06,545 WARN  [master:60000.activeMasterManager] master.SplitLogManager: returning success without actually splitting and deleting all the log files in path hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting
2017-11-25 10:38:06,545 INFO  [master:60000.activeMasterManager] master.SplitLogManager: finished splitting (more than or equal to) 83 bytes in 1 log files in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] in 11690ms
2017-11-25 10:38:06,929 INFO  [master:60000.activeMasterManager] zookeeper.MetaTableLocator: Failed verification of hbase:meta,,1 at address=slave2,16020,1511532219830, exception=org.apache.hadoop.hbase.NotServingRegionException: Region hbase:meta,,1 is not online on slave2,16020,1511577423313
at org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2915)
at org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:979)
at org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegionInfo(RSRpcServices.java:1259)
at org.apache.hadoop.hbase.protobuf.generated.AdminProtos$AdminService$2.callBlockingMethod(AdminProtos.java:22233)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2145)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107)
at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:748)


2017-11-25 10:38:06,941 INFO  [master:60000.activeMasterManager] master.SplitLogManager: dead splitlog workers [slave2,16020,1511532219830]
2017-11-25 10:38:06,951 INFO  [master:60000.activeMasterManager] master.SplitLogManager: hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting is empty dir, no logs to split
2017-11-25 10:38:06,951 INFO  [master:60000.activeMasterManager] master.SplitLogManager: started splitting 0 logs in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] for [slave2,16020,1511532219830]
2017-11-25 10:38:06,977 WARN  [master:60000.activeMasterManager] master.SplitLogManager: returning success without actually splitting and deleting all the log files in path hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting
2017-11-25 10:38:06,977 INFO  [master:60000.activeMasterManager] master.SplitLogManager: finished splitting (more than or equal to) 0 bytes in 0 log files in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] in 26ms
2017-11-25 10:38:06,978 INFO  [master:60000.activeMasterManager] zookeeper.MetaTableLocator: Deleting hbase:meta region location in ZooKeeper
2017-11-25 10:38:07,292 INFO  [master:60000.activeMasterManager] master.AssignmentManager: Setting node as OFFLINED in ZooKeeper for region {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}
2017-11-25 10:38:07,385 INFO  [master:60000.activeMasterManager] master.AssignmentManager: Assigning hbase:meta,,1.1588230740 to slave2,16020,1511577423313
2017-11-25 10:38:07,385 INFO  [master:60000.activeMasterManager] master.RegionStates: Transition {1588230740 state=OFFLINE, ts=1511577487292, server=null} to {1588230740 state=PENDING_OPEN, ts=1511577487385, server=slave2,16020,1511577423313}
2017-11-25 10:38:07,595 INFO  [master:60000.activeMasterManager] master.ServerManager: AssignmentManager hasn't finished failover cleanup; waiting
2017-11-25 10:38:07,837 INFO  [AM.ZK.Worker-pool2-t1] master.RegionStates: Transition {1588230740 state=PENDING_OPEN, ts=1511577487385, server=slave2,16020,1511577423313} to {1588230740 state=OPENING, ts=1511577487837, server=slave2,16020,1511577423313}
2017-11-25 10:38:12,722 INFO  [AM.ZK.Worker-pool2-t2] master.RegionStates: Transition {1588230740 state=OPENING, ts=1511577487837, server=slave2,16020,1511577423313} to {1588230740 state=OPEN, ts=1511577492722, server=slave2,16020,1511577423313}
2017-11-25 10:38:12,729 INFO  [AM.ZK.Worker-pool2-t2] coordination.ZkOpenRegionCoordination: Handling OPENED of 1588230740 from master,60000,1511577430262; deleting unassigned node
2017-11-25 10:38:12,882 INFO  [master:60000.activeMasterManager] master.HMaster: hbase:meta with replicaId 0 assigned=1, rit=false, location=slave2,16020,1511577423313
2017-11-25 10:38:13,691 INFO  [master:60000.activeMasterManager] hbase.MetaMigrationConvertingToPB: META already up-to date with PB serialization
2017-11-25 10:38:13,878 INFO  [master:60000.activeMasterManager] master.RegionStates: Transition {e6bff28294ecdbdc5fadca964ceafb26 state=OPEN, ts=1511577493878, server=slave1,16020,1511532206980} to {e6bff28294ecdbdc5fadca964ceafb26 state=OFFLINE, ts=1511577493878, server=slave1,16020,1511532206980}
2017-11-25 10:38:13,899 INFO  [master:60000.activeMasterManager] master.RegionStates: Transition {1d3b7b34795a72f8fb15b68d64c56e53 state=OPEN, ts=1511577493899, server=master,16020,1511532189674} to {1d3b7b34795a72f8fb15b68d64c56e53 state=OFFLINE, ts=1511577493899, server=master,16020,1511532189674}
2017-11-25 10:38:13,994 INFO  [master:60000.activeMasterManager] master.AssignmentManager: Found regions out on cluster or in RIT; presuming failover
2017-11-25 10:38:14,105 INFO  [master:60000.activeMasterManager] master.AssignmentManager: Joined the cluster in 414ms, failover=true
2017-11-25 10:38:14,334 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] handler.ServerShutdownHandler: Splitting logs for slave1,16020,1511532206980 before assignment; region count=1
2017-11-25 10:38:14,367 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] handler.ServerShutdownHandler: Splitting logs for slave2,16020,1511532219830 before assignment; region count=0
2017-11-25 10:38:14,424 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] handler.ServerShutdownHandler: Splitting logs for master,16020,1511532189674 before assignment; region count=1
2017-11-25 10:38:14,483 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] master.SplitLogManager: dead splitlog workers [slave1,16020,1511532206980]
2017-11-25 10:38:14,506 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] master.SplitLogManager: started splitting 1 logs in [hdfs://master:9000/hbase/WALs/slave1,16020,1511532206980-splitting] for [slave1,16020,1511532206980]
2017-11-25 10:38:14,559 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] master.SplitLogManager: dead splitlog workers [slave2,16020,1511532219830]
2017-11-25 10:38:14,617 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] master.SplitLogManager: started splitting 1 logs in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] for [slave2,16020,1511532219830]
2017-11-25 10:38:14,655 INFO  [master,60000,1511577430262_splitLogManager__ChoreService_1] master.SplitLogManager: total tasks = 2 unassigned = 2 tasks={/hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830.default.1511532372089=last_update = -1 last_version = -1 cur_worker_name = null status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0, /hbase/splitWAL/WALs%2Fslave1%2C16020%2C1511532206980-splitting%2Fslave1%252C16020%252C1511532206980.default.1511532378995=last_update = 1511577494615 last_version = -1 cur_worker_name = null status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0}
2017-11-25 10:38:14,693 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave1%2C16020%2C1511532206980-splitting%2Fslave1%252C16020%252C1511532206980.default.1511532378995 acquired by slave2,16020,1511577423313
2017-11-25 10:38:14,818 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] master.SplitLogManager: dead splitlog workers [master,16020,1511532189674]
2017-11-25 10:38:14,907 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] master.SplitLogManager: started splitting 1 logs in [hdfs://master:9000/hbase/WALs/master,16020,1511532189674-splitting] for [master,16020,1511532189674]
2017-11-25 10:38:15,171 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830.default.1511532372089 acquired by master,16020,1511577430553
2017-11-25 10:38:15,337 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fmaster%2C16020%2C1511532189674-splitting%2Fmaster%252C16020%252C1511532189674.default.1511532381539 acquired by slave1,16020,1511577422976
2017-11-25 10:38:20,646 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave1%2C16020%2C1511532206980-splitting%2Fslave1%252C16020%252C1511532206980.default.1511532378995 entered state: DONE slave2,16020,1511577423313
2017-11-25 10:38:20,654 INFO  [master,60000,1511577430262_splitLogManager__ChoreService_1] master.SplitLogManager: total tasks = 3 unassigned = 0 tasks={/hbase/splitWAL/WALs%2Fmaster%2C16020%2C1511532189674-splitting%2Fmaster%252C16020%252C1511532189674.default.1511532381539=last_update = 1511577496132 last_version = 2 cur_worker_name = slave1,16020,1511577422976 status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0, /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830.default.1511532372089=last_update = 1511577495619 last_version = 2 cur_worker_name = master,16020,1511577430553 status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0, /hbase/splitWAL/WALs%2Fslave1%2C16020%2C1511532206980-splitting%2Fslave1%252C16020%252C1511532206980.default.1511532378995=last_update = 1511577500561 last_version = 2 cur_worker_name = slave2,16020,1511577423313 status = in_progress incarnation = 0 resubmits = 0 batch = installed = 1 done = 0 error = 0}
2017-11-25 10:38:20,905 INFO  [main-EventThread] wal.WALSplitter: Archived processed log hdfs://master:9000/hbase/WALs/slave1,16020,1511532206980-splitting/slave1%2C16020%2C1511532206980.default.1511532378995 to hdfs://master:9000/hbase/oldWALs/slave1%2C16020%2C1511532206980.default.1511532378995
2017-11-25 10:38:20,909 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: Done splitting /hbase/splitWAL/WALs%2Fslave1%2C16020%2C1511532206980-splitting%2Fslave1%252C16020%252C1511532206980.default.1511532378995
2017-11-25 10:38:20,920 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] master.SplitLogManager: finished splitting (more than or equal to) 83 bytes in 1 log files in [hdfs://master:9000/hbase/WALs/slave1,16020,1511532206980-splitting] in 6414ms
2017-11-25 10:38:20,927 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] handler.ServerShutdownHandler: Reassigning 1 region(s) that slave1,16020,1511532206980 was carrying (and 0 regions(s) that were opening on this server)
2017-11-25 10:38:20,928 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] handler.ServerShutdownHandler: Best effort in SSH to retain assignment of 1 regions from the dead server slave1,16020,1511532206980
2017-11-25 10:38:21,241 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830.default.1511532372089 entered state: DONE master,16020,1511577430553
2017-11-25 10:38:21,262 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] balancer.BaseLoadBalancer: Reassigned 1 regions. 1 retained the pre-restart assignment. 
2017-11-25 10:38:21,263 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] master.AssignmentManager: Bulk assigning 1 region(s) across 3 server(s), retainAssignment=true
2017-11-25 10:38:21,332 INFO  [main-EventThread] wal.WALSplitter: Archived processed log hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting/slave2%2C16020%2C1511532219830.default.1511532372089 to hdfs://master:9000/hbase/oldWALs/slave2%2C16020%2C1511532219830.default.1511532372089
2017-11-25 10:38:21,342 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: Done splitting /hbase/splitWAL/WALs%2Fslave2%2C16020%2C1511532219830-splitting%2Fslave2%252C16020%252C1511532219830.default.1511532372089
2017-11-25 10:38:21,361 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] master.SplitLogManager: finished splitting (more than or equal to) 83 bytes in 1 log files in [hdfs://master:9000/hbase/WALs/slave2,16020,1511532219830-splitting] in 6744ms
2017-11-25 10:38:21,362 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] handler.ServerShutdownHandler: Reassigning 0 region(s) that slave2,16020,1511532219830 was carrying (and 0 regions(s) that were opening on this server)
2017-11-25 10:38:21,362 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] handler.ServerShutdownHandler: Best effort in SSH to retain assignment of 0 regions from the dead server slave2,16020,1511532219830
2017-11-25 10:38:21,363 INFO  [MASTER_SERVER_OPERATIONS-master:60000-0] handler.ServerShutdownHandler: Finished processing of shutdown of slave2,16020,1511532219830
2017-11-25 10:38:21,424 INFO  [master,60000,1511577430262-GeneralBulkAssigner-1] master.AssignmentManager: Assigning 1 region(s) to slave1,16020,1511577422976
2017-11-25 10:38:21,631 INFO  [master,60000,1511577430262-GeneralBulkAssigner-1] master.RegionStates: Transition {e6bff28294ecdbdc5fadca964ceafb26 state=OFFLINE, ts=1511577501524, server=slave1,16020,1511532206980} to {e6bff28294ecdbdc5fadca964ceafb26 state=PENDING_OPEN, ts=1511577501631, server=slave1,16020,1511577422976}
2017-11-25 10:38:22,777 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] master.AssignmentManager: Bulk assigning done
2017-11-25 10:38:22,777 INFO  [MASTER_SERVER_OPERATIONS-master:60000-1] handler.ServerShutdownHandler: Finished processing of shutdown of slave1,16020,1511532206980
2017-11-25 10:38:23,054 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: task /hbase/splitWAL/WALs%2Fmaster%2C16020%2C1511532189674-splitting%2Fmaster%252C16020%252C1511532189674.default.1511532381539 entered state: DONE slave1,16020,1511577422976
2017-11-25 10:38:23,068 INFO  [AM.ZK.Worker-pool2-t5] master.RegionStates: Transition {e6bff28294ecdbdc5fadca964ceafb26 state=PENDING_OPEN, ts=1511577501631, server=slave1,16020,1511577422976} to {e6bff28294ecdbdc5fadca964ceafb26 state=OPENING, ts=1511577503068, server=slave1,16020,1511577422976}
2017-11-25 10:38:23,137 INFO  [main-EventThread] wal.WALSplitter: Archived processed log hdfs://master:9000/hbase/WALs/master,16020,1511532189674-splitting/master%2C16020%2C1511532189674.default.1511532381539 to hdfs://master:9000/hbase/oldWALs/master%2C16020%2C1511532189674.default.1511532381539
2017-11-25 10:38:23,142 INFO  [main-EventThread] coordination.SplitLogManagerCoordination: Done splitting /hbase/splitWAL/WALs%2Fmaster%2C16020%2C1511532189674-splitting%2Fmaster%252C16020%252C1511532189674.default.1511532381539
2017-11-25 10:38:23,177 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] master.SplitLogManager: finished splitting (more than or equal to) 83 bytes in 1 log files in [hdfs://master:9000/hbase/WALs/master,16020,1511532189674-splitting] in 8270ms
2017-11-25 10:38:23,180 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] handler.ServerShutdownHandler: Reassigning 1 region(s) that master,16020,1511532189674 was carrying (and 0 regions(s) that were opening on this server)
2017-11-25 10:38:23,181 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] handler.ServerShutdownHandler: Best effort in SSH to retain assignment of 1 regions from the dead server master,16020,1511532189674
2017-11-25 10:38:23,245 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] balancer.BaseLoadBalancer: Reassigned 1 regions. 1 retained the pre-restart assignment. 
2017-11-25 10:38:23,258 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] master.AssignmentManager: Bulk assigning 1 region(s) across 3 server(s), retainAssignment=true
2017-11-25 10:38:23,284 INFO  [master,60000,1511577430262-GeneralBulkAssigner-0] master.AssignmentManager: Assigning 1 region(s) to master,16020,1511577430553
2017-11-25 10:38:23,395 INFO  [master,60000,1511577430262-GeneralBulkAssigner-0] master.RegionStates: Transition {1d3b7b34795a72f8fb15b68d64c56e53 state=OFFLINE, ts=1511577503339, server=master,16020,1511532189674} to {1d3b7b34795a72f8fb15b68d64c56e53 state=PENDING_OPEN, ts=1511577503395, server=master,16020,1511577430553}
2017-11-25 10:38:24,489 INFO  [AM.ZK.Worker-pool2-t7] master.RegionStates: Transition {1d3b7b34795a72f8fb15b68d64c56e53 state=PENDING_OPEN, ts=1511577503395, server=master,16020,1511577430553} to {1d3b7b34795a72f8fb15b68d64c56e53 state=OPENING, ts=1511577504489, server=master,16020,1511577430553}
2017-11-25 10:38:29,242 INFO  [AM.ZK.Worker-pool2-t8] master.RegionStates: Transition {e6bff28294ecdbdc5fadca964ceafb26 state=OPENING, ts=1511577503068, server=slave1,16020,1511577422976} to {e6bff28294ecdbdc5fadca964ceafb26 state=OPEN, ts=1511577509242, server=slave1,16020,1511577422976}
2017-11-25 10:38:29,372 INFO  [AM.ZK.Worker-pool2-t10] master.RegionStates: Offlined e6bff28294ecdbdc5fadca964ceafb26 from slave1,16020,1511532206980
2017-11-25 10:38:34,136 INFO  [master:60000.activeMasterManager] master.HMaster: Master has completed initialization
2017-11-25 10:38:34,219 INFO  [master:60000.activeMasterManager] quotas.MasterQuotaManager: Quota support disabled
2017-11-25 10:38:34,264 INFO  [master:60000.activeMasterManager] zookeeper.ZooKeeperWatcher: not a secure deployment, proceeding
2017-11-25 10:38:34,891 INFO  [AM.ZK.Worker-pool2-t11] master.RegionStates: Transition {1d3b7b34795a72f8fb15b68d64c56e53 state=OPENING, ts=1511577504489, server=master,16020,1511577430553} to {1d3b7b34795a72f8fb15b68d64c56e53 state=OPEN, ts=1511577514891, server=master,16020,1511577430553}
2017-11-25 10:38:34,915 INFO  [AM.ZK.Worker-pool2-t13] master.RegionStates: Offlined 1d3b7b34795a72f8fb15b68d64c56e53 from master,16020,1511532189674
2017-11-25 10:38:34,917 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] master.AssignmentManager: Bulk assigning done
2017-11-25 10:38:34,917 INFO  [MASTER_SERVER_OPERATIONS-master:60000-2] handler.ServerShutdownHandler: Finished processing of shutdown of master,16020,1511532189674

二、在網上查看了很多的介紹,感覺與我的情況不太相符

  於是我查看了叢集中的程序資訊發現zookeeper叢集的的程序沒有了,重啟zookeeper叢集,在啟動HBASE就行

使用netstat  -nltp 命令檢視埠資訊


60010埠已經正常了,問題解決。

相關推薦

HBASE-1.1.12叢集啟動之後HmasterHregionServer程序很快問題

一、檢視日誌資訊如下   2017-11-25 10:37:25,646 INFO  [main] zookeeper.ZooKeeper: Client environment:java.library.path=/usr/local/hadoop-2.7.4/lib/n

hadoop叢集啟動之後datanode程序啟動解決辦法

這種情況一般是由於自己進行hadoop格式化的時候沒有事先結束所有程序,或者在開啟一次hadoop叢集,發現配置不對,這時重新 hadoop namenode -format  就會出問題 因為此時你format之後, datanode的clusterID 和 namenode的clu

啟動hbaseHmasterHRegionServer程序有時候起不來

它出現的原因可能有是以下兩點: 第一,可能是各個主機的時間不同步 【1】在各個主機上執行:date 命令,看看時間是否同步 【2】如果時間不同步的話 格式如下: #  ntpdate      -u&nbs

1天 | 12天搞定Python告訴你有什麼用?

掌握多一門程式語言,多一種選擇,多一份機遇,更何況學的是人見人愛,花見花開的Python語言。它目前可佔據程式語言排行榜的第3名,是名副其實的“探花郎”,無論用它做什麼(網路爬蟲、人工智慧、Web應用等) ,咱都不虧,開篇有益。 有趣的話,多說兩句,”無趣”的就總結概述。本章總結概述一下Python的發展歷史

Ubuntu和Win7雙系統ubuntu被刪重新啟動之後顯示no such partition

div cmd命令 class ubuntu 終端 oot windows7 data- 重新啟動 準備一張windows7的系統安裝盤。從光盤啟動電腦,在光盤啟動完畢之後。按shift+F10,調出cmd命令終端,在終端輸入:bootrec/fixmbr OK。重新啟

叢集啟動使用Hadoop執行後沒有NameNode節點

叢集啟動Hadoop後,執行後有SecondNameNode和dataNode,但是沒有NameNode節點? 解決方法: 1、先執行stop-all.sh 2、執行hadoop namenode

程序退出之後程序發生什麼?

在linux中,所有程序都有一個共同的父程序systemd,如果父程序退出了,子程序還沒執行結束,子程序會被stsremd收養 下面用一個小程式來驗證一下: #include <cstdio> #include <sys/types.h> #inc

zookeepe啟動顯示成功但是程序沒有(實際沒有啟動成功)

使用命令/zookeeper/bin/zkServer.sh start啟動zookeeper 顯示Starting zookeeper 。。。 STARTED  但是通過ps命令和telnet localhost 2181發現無服務啟動 通過echo $JAVA_HOME檢視,

kube-apiserver進程無法啟動k8s 1.7.12

k8s kubernetes kube-apiserver 服務 無法啟動 [root@k8s_m3:~]# journalctl -fu kube-apiserver-- Logs begin at Fri 2018-03-23 05:28:27 CST. --Mar 23 11:21:

Hadoop ha CDH5.15.1-hadoop叢集啟動兩個namenode都是standby模式

         Hadoop ha CDH5.15.1-hadoop叢集啟動後,兩個namenode都是standby模式                                              作者:尹正傑 版權宣告:原創作品,謝絕轉載!否則將追究法律責任。   &nb

使用開源Breeze工具部署Kubernetes 1.12.1高可用叢集_Kubernetes中文社群

Breeze專案是深圳睿雲智合所開源的Kubernetes圖形化部署工具,大大簡化了Kubernetes部署的步驟,其最大亮點在於支援全離線環境的部署,且不需要翻牆獲取Google的相應資源包,尤其適合某些不便訪問網際網路的伺服器場景。(專案地址 https://github.com/wise

用kubeadm在centos7下搭建k8s 1.12叢集

在搭建的過程中遇到一個難題 本地將下載的映象包,匯入到本地並打好tag後,執行 kubeadm init命令,還會一直去拉取映象,不知道原因? [[email protected] log]# kubeadm  init I1113 15:31:40.17659

解決 win 8.1 + Ubuntu 12.04 安裝雙系統後啟動無法進入ubuntu的問題

本著不折騰不舒服的原則,今天將一塊嶄新的SSD安裝了win8.1 + ubuntu 12.04的雙系統。之前win8+ubuntu的組合不知道裝了多少遍了,沒想到今天裝完win8.1 + ubuntu啟動後,卻不能進入grub介面,而是直接進入win8.1。 前置條件

MongoDB分片叢集啟動過程中Failed to connect to 127.0.0.1:27017錯誤解決辦法

(1)進行MongoDB分片叢集部署過後,進入MongoDB的目錄下,輸入bin/mongo -port:30000,居然彈出錯誤: MongoDB shell version: 3.0.7 connecting to: test 2015-10-30T10:34:27.2

安裝12.1.0.2 叢集GRID/GI, 執行root.sh 指令碼失敗的案例

在Linux系統上安裝12.1.0.2 叢集GRID/GI軟體,節點2執行root.sh失敗,螢幕的錯誤資訊: OLR initialization - successful 2015/12/15 13:16:55 CLSRSC-507: The root script c

Hadoop-2.6.0+Zookeeper-3.4.6+Spark-1.5.0+Hbase-1.1.2+Hive-1.2.0叢集搭建

前言 本部落格目的在於跟大家分享大資料平臺搭建過程,是筆者半年的結晶。在大資料搭建過程中,希望能給大家提過一些幫助,這也是本部落格的

搭建kafka2.12-1.0.0+storm1.1.1叢集

各應用程式及版本CentOS65jdk1.8.0_121storm1.1.1zookeeper3.4.10kafka2.12-1.0.0redis4.0.5storm、zookeeper、kafka可從apache.org網站獲取相應版本,建議版本與本文件一致;否則後續的示例

(13)HBase 1.2.6 叢集部署

1.引言       在前面的部落格中我們介紹HBase的偽分佈安裝,在本篇部落格中就主要介紹一下HBase的叢集部署,由於HBase是一款Hadoop之上的NoSQL資料庫,因此我們在安裝

cool-2018-10-22-centos7-hive-1.2叢集+整合hbase-1.1.3叢集

hive與hbase整合版本相容說明 HIVE 0.9.0,HBASE整合至少需要HBASE 0.92,HIVE的早期版本與HBase 0.89/0.90一起使用。 hive1.X將保持與HBase 0.98.x和更低版本相容。hive2x將與HBASE 1 x

2017.7.1 mysql安裝與啟動(已驗證可以使用)

界面 all avi 配置 install 服務 毫無 壓力 cnblogs 之前一直用解壓版安裝,啟動mysql服務的時候總是失敗,這次用mysql installer安裝一遍,終於成功啟動。 1.下載mysql installer 下載的32位的,在64位系統運行也