版權(quán)說明:本文檔由用戶提供并上傳,收益歸屬內(nèi)容提供方,若內(nèi)容存在侵權(quán),請進(jìn)行舉報或認(rèn)領(lǐng)
文檔簡介
Information
Change
t1heWorldBONC提綱—Hadoop集群架構(gòu)三Hadoop安裝部署二Hadoop日常操作與四Hadoop高可用相關(guān)Information
Change
the
WorldBONCHadoop2.3主要進(jìn)程Hadoop基礎(chǔ)進(jìn)程高可用相關(guān)進(jìn)程NamenodeDataNodeResourceManagerNodeManagerQuorumPeerMainZKFailoverControllerJournalNodeZookeeper為ZKFC和MRZKFC提供了基礎(chǔ)服務(wù).保證了
ameNode節(jié)點服務(wù)級別的自動熱切.JournalNode保證了NameNode雙機元數(shù)據(jù)的強一致性.BONCJournalNodeJournalNodeJournalNodeJournalNodeJournalNodeNameNodeActiveNameNodeStandby寫入Active狀態(tài)的NameNode負(fù)責(zé)響應(yīng)所有的客戶端操作,并將相關(guān)操作對EditLog所做 寫入到所有的JournalNode.Standby狀況的NameNode才能 JournalNode中的變動信息,并且Edit
Log的更改應(yīng)用于本身的元數(shù)據(jù)信息.QJM的基本原理就是用2N+1臺JournalNodeEditLog,每次寫數(shù)據(jù)操作有大多數(shù)(>=N+1)返回成功時即認(rèn)為該次寫成功,數(shù)據(jù)不的是最多有N臺機器掛掉,如果多于會丟失了.當(dāng)然這個算法所能N臺掛掉,這個算法就失效了.QJM/Qurom
Journal
ManagerInformation
Change
t3he
WorldBONC提綱—Hadoop集群架構(gòu)三Hadoop安裝部署二Hadoop日常操作與四Hadoop高可用相關(guān)Information
Change
t4he
WorldBONC集群架構(gòu)拓?fù)鋱Dhadoop001hadoop003hadoop005hadoop007hadoop002hadoop004hadoop006hadoop008hadoop009hadoop010Information
Change
t5he
WorldDataNodeNameNodeMysql交換機以10臺hadoop服務(wù)器為例來講解一下安裝實施的過程步驟.集群共有10臺服務(wù)器,分別為hadoop001至hadoop010安裝hadoop組件包括:HDFS、Mapreduce、Yarn、ZooKeeper、Hive、Sqoop等BONC集群各節(jié)點服務(wù)hadoop001和hadoop002作為NameNode互為熱備。hadoop003、hadoop004、hadoop005、hadoop006、hadoop007、hadoop008作為DataNode節(jié)點存放數(shù)據(jù)。Hive元數(shù)據(jù)庫Mysql安裝在hadoop009和hadoop010做主從
。相關(guān)ETL流程及其他應(yīng)用部署hadoop009和hadoop010。Information
Change
t6he
WorldBONC安裝準(zhǔn)備工作修改/etc/hosts添加所有節(jié)點及相應(yīng)主機名并分發(fā).添加hadoop用戶并配置hadoop用戶之間的ssh互信(免密).關(guān)閉 和Selinux.配置時間同步ntp服務(wù).創(chuàng)建DataNode磁盤數(shù)據(jù) 用以存放數(shù)據(jù),并修改為hadoop屬主./opt/beh 修改為hadoop屬主所有操作修改以及啟停服務(wù)在hadoop用戶下進(jìn)行將 tar包beh.tar解壓至/opt
下修改相關(guān)配置文件并且分發(fā)至所有集群節(jié)點Information
Change
the
WorldInformation
Change
the
WorldBONCZooKeeper相關(guān)配置文件/opt/beh/core/zookeeper/conf/zoo.cfg#
The
number
of
milliseconds
of
each
ticktickTime=2000#
The
number
of
ticks
that
the
initial#
synchronization
phase
cantakeinitLimit=10#
The
number
of
ticks
that
can
pass
between#
sending
a
request
and
getting
an
acknowledgementsyncLimit=5#
the
directorywhere
the
snapshot
is
stored.#
do
not
use
/tmp
for
storage,
/tmp
here
is
just#
example
sakes.dataDir=/opt/beh/zookeeper/data#
the
portat
whichthe
clients
will
connectclientPort=2181server.1=hadoop001:2888:3888server.2=hadoop002:2888:3888server.3=hadoop003:2888:3888##
Be
sure
to
read
the
maintenance
section
ofthe#
administrat uide
before
turning
on
autopurge.####
The
number
o apshots
to
retain
in
dataDir#autopurge.snapRetainCount=3#
Purge
task
interval
in
hours#
Set
to
"0"
to
disable
auto
purge
feature#autopurge.purgeInterval=1配置文件中的dataDir參數(shù)對應(yīng)的各zookeeper節(jié)點的唯一性標(biāo)示.在相應(yīng) 下創(chuàng)建文件myid,分別填寫相應(yīng)的邏輯id.Information
Change
the
WorldBONCHDFS配置相關(guān)文件core-site.xml<configuration><property><name>fs.defaultFS</name><value>hdfs://beh</value></property><property><name>hadoop.tmp.dir</name><value>/opt/beh/data/namenode</value></property><property><name>ha.zookeeper.quorum</name><value>hadoop001:2181,hadoop002:2181,hadoop003:2181</value></property><property><name>io.file.buffer.size</name><value>131072</value></property></configuration>/opt/beh/hadoop/etc/hadoop/core-site.xmlcore-site.xmlInformation
Change
the
WorldBONCHDFS配置相關(guān)文件hdfs-site.xml<property><name>dfs.data.dir</name><value>/data/disk1/data</value></property><property><name>dfs.block.size</name><value>134217728</value></property><name>dfs.datanode.failed.volumes.tolerated</name><value>4</value></property><property><name>dfs.replication</name><value>3</value></property>/opt/beh/core/hadoop/etc/hadoop/hdfs-site.xml/opt/beh/core/hadoop/etc/hadoop/slave填寫所有DataNode節(jié)點的主機名或者IP地址.hadoop003hadoop004hadoop005hadoop006hadoop007hadoop008hdfs-site.xmlBONCMapred配置相關(guān)文件mapred-site.xml<property><name></name><value>yarn</value></property><property><name>mapreduce.jobhistory.address</name><value>hadoop001:10020</value></property><property><name>mapreduce.jobhistory.webapp.address</name><value>hadoop001:19888</value></property><property><name>mapred.child.java.opts</name><value>-Xmx2g</value></property>/opt/beh/core/hadoop/etc/hadoop/mapred-site.xmlmapred-site.xmlInformation
Change
the
WorldBONCYARN配置相關(guān)文件yarn-site.xml<property><name>yarn.resourcemanager.cluster-id</name><value>beh</value></property><property><name>yarn.resourcemanager.ha.rm-ids</name><value>rm1,rm2</value></property><property><name>yarn.resourcemanager.ha.id</name><value>rm1</value></property>/opt/beh/core/hadoop/etc/hadoop/yarn-site.xmlarn-site.xmlInformation
Change
the
WorldBONCHbase配置相關(guān)文件hbase-site.xml<property><name>hbase.zookeeper.quorum</name><value>hadoop001,hadoop002,hadoop003</value></property><property><name>dfs.replication</name><value>3</value></property><property><name>hbase.hregion.max.filesize</name><value>2147483648</value></property>/opt/beh/core/hbase/conf/hbase-site.xml/opt/beh/core/hbase/conf/hbase-site.xml/regionservershadoop003hadoop004hadoop005hadoop006hadoop007hadoop008填寫所有需要啟動Regionserver節(jié)點的主機名或者IP地址.hbase-site.xmlInformation
Change
the
WorldBONCHive相關(guān)配置文件hive-site.xml<property><name>javax.jdo.option.ConnectionURL</name><value>jdbc:mysql://hivemysql:3306/hive?createDatabaseIfNotExist=true;char;useUnicode=true</value>acterEncoding=UTF-8</property><property><name>javax.jdo.option.ConnectionDriverName</name><value>com.mysql.jdbc.Driver</value></property><property><name>javax.jdo.option.ConnectionUserName</name><value>hadoop</value></property><property><name>javax.jdo.option.ConnectionPassword</name><value>hadoop</value></property>Information
Change
the
WorldInformation
Change
the
WorldBONC環(huán)境變量相關(guān)配置ME/bin:$SPARK_HOME/bin:$MAHOUT_HOME/bin:$PATH集群所有節(jié)點修改/etc/profile添加如下內(nèi)容exportBEH_HOME=/opt/behexportJAVA_HOME=/opt/beh/core/jdk1.7.0_55exportHBASE_HOME=/opt/beh/core/hbaseexportHIVE_HOME=/opt/beh/core/hiveexport
HADOOP_HOME=/opt/beh/core/hadoopexportMAHOUT_HOME=/opt/beh/core/mahoutexport
MON_HOME=/opt/beh/core/hadoopexportHADOOP_MAPRED_HOME=/opt/beh/core/hadoopexportSQOOP_HOME=/opt/beh/core/sqoopexportZOOKEEPER_HOME=/opt/beh/core/zookeeper
exportLD_LIBRARY_PATH=/opt/beh/core/hadoop/lib/nativeexportPUB_KEY_HOME=/home/hadoop/.ssh/id_rsaexportZOO_LOG_DIR=$BEH_HOME/logs/zookeeperexport
SCALA_HOME=/opt/beh/core/scala-2.9.3exportSPARK_HOME=/opt/beh/core/sparkexport
SPRAK_EXAMPLES_JAR=/opt/beh/core/spark/examples/lib/spark-examples_2.10-0.9.0-cdh5.0.0.jarexportCLASSPATH=.:$JAVA_HOME/lib:$JAVA_HOME/jre/lib:$SQOOP_HOME/lib:$HIVE_HOME/lib:$HADOOP_HOME/lib:/opt/beh/core/hive/hcatalog/share/hcatalog:$CLASSPATHexportPATH=$SCALA_HOME/bin:$SPARK_HOME/bin:$ZOOKEEPER_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$JAVA_HOME/bin:$JAVA_HOME/jre/bin:$HBASE_HOME/bin:$HIVE_HOME/bin:$SQOOP_HOBONCInformation
Change
the
World啟動并格式化ZooKeeper集群所有zookeeper節(jié)點分別執(zhí)行:zkServer.sh
start驗證:zkCli.sh
并執(zhí)行l(wèi)s/再執(zhí)行quit可推出交互.執(zhí)行jps命令檢測是否存在zookeeper服務(wù)在hadoop001上執(zhí)行:hdfs
zkfc-formatZK驗證:zkCli.sh
并執(zhí)行l(wèi)s/再執(zhí)行quit可推出交互執(zhí)行如下命令在所有zookeeper服務(wù)器查看zookeeper角色zkServer.sh
statusBONC啟動并格式化NameNode所有journalnode分別執(zhí)行:/opt/beh/core/hadoop/sbin/hadoop-daemon.sh
start
journalnodejps檢測是否存在journalnode服務(wù)任選一個,這里選擇hadoop001,執(zhí)行:hd amenode
-formathd amenode
-initializeSharedEditshadoop-daemon.sh
start
namenodejps驗證namenode服務(wù)在hadoop002執(zhí)行:hd amenode
-bootstrapStandby在hadoop002執(zhí)行:hadoop-daemon.sh
start
namenode執(zhí)行jps驗證namenode服務(wù)是否存在皆顯示此時登陸
與Standby狀態(tài)Information
Change
the
WorldBONC啟動并格式化DataNode在hadoop001執(zhí)行(這個是 全啟動):hadoop-daemons.sh
start
datanode或者在每個datanode上執(zhí)行hadoop-daemon.sh
start
datanodejps驗證是否存在datanode服務(wù)注:對于重建的Hadoop集群,格式化NameNode時并不會將DataNode的舊文件刪除,需要手動將數(shù)據(jù)和臨時文件 下的舊文件刪除.否則造成DataNode啟動失敗.Information
Change
the
WorldBONC啟動并查看
ameNode狀態(tài)所有NameNode節(jié)點分別執(zhí)行:hadoop-daemon.sh
start
zkfc驗證:zkCli.sh
并執(zhí)行l(wèi)s/再執(zhí)行quit可推出交互登陸namenode服務(wù)器web端查看服務(wù)器狀態(tài)與其中一個為active另一個為此時登陸standby狀態(tài)Information
Change
the
WorldBONCInformation
Change
the
World啟動YARN在hadoop001,hadoop002上啟動:yarn-daemon.sh
start
resourcemanager在hadoop003,hadoop004,hadoop005.....hadoop008上啟動:yarn-daemon.sh
start
datamanagerBONCInformation
Change
the
World啟動Hbase相關(guān)服務(wù)啟動regionserver#在所有regionserver角色上執(zhí)行
hbase-daemon.sh
start
regionserver啟動hmaster#在需要登陸的客戶端上執(zhí)行
hbase-daemon.sh
start
masterBONCHive元數(shù)據(jù)Mysql數(shù)據(jù)庫設(shè)置安裝Mysql數(shù)據(jù)庫,一般選擇操作系統(tǒng)安裝包自帶的rpm包安裝即可.MySQL安裝及主從配置.txtInformation
Change
the
WorldBONCInformation
Change
the
WorldHive元數(shù)據(jù)Mysql庫字符集問題目前在Hive中使用中文注釋時,會有部分問題.desc
tablename查看注釋可通過下方法解決.但是show
create
table
tablename仍然存在問題.解決辦法:mysql
-uroot
-pmysql123usehive;show
tables;
##找到COLUMNS的那張表,表名因版本略有不同,看下數(shù)據(jù)。show
full
columnsfrom
COLUMNS_V2;
##表名是大小寫敏感的。
alter
table
COLUMNS_V2
modify
column
COMMENT
varchar(256)character
set
utf8;show
full
columns
from
TABLE_PARAMS;alter
table
TABLE_PARAMSmodify
column
PARAM_VALUEvarchar(4000)
character
set
utf8;BONC運行Hive與jar包程序使用hive使用hive客戶端時應(yīng)使用hive命令.Debug模式,如果有什么報錯可以用這個進(jìn) 試查看問題在哪里./hive
-hiveconfhive.root.logger=DEBUG,console運行jar包程序在運行jar包程序時需使用/opt/beh/core/hadoop/bin-mapreduce1目
錄下hadoop相關(guān)命令I(lǐng)nformation
Change
the
WorldBONC提綱—Hadoop集群架構(gòu)三Hadoop安裝部署二Hadoop日常操作與四Hadoop高可用相關(guān)Information
Change
t2h5eWorldBONCInformation
Change
t2h6eWorld單服務(wù)啟動命令(一)啟動zookeeper服務(wù)所有zookeeper服務(wù)節(jié)點執(zhí)行:zkServer.sh
start啟動JournalNode服務(wù)所有journalenode節(jié)點執(zhí)行:hadoop-daemon.sh
start
journalnode啟動NameNode服務(wù)所有NameNode節(jié)點執(zhí)行hadoop-daemon.sh
start
namenode啟動DataNode服務(wù)在某一個節(jié)點上執(zhí)行:hadoop-daemons.sh
start
datanode或者在所有DataNode節(jié)點上執(zhí)行:(某些時候需要單獨啟動某個節(jié)點的服務(wù))hadoop-daemon.sh
start
datanodeBONCInformation
Change
t2h7eWorld單服務(wù)啟動命令(二)啟動ZooKeeperFailoverController在所有NameNode節(jié)點上執(zhí)行:
hadoop-daemon.sh
start
zkfc啟動ResourceManager在所有resoucermanager節(jié)點上執(zhí)行:
yarn-daemon.sh
start
resoucermanager啟動nodemanager在所有nodemanager節(jié)點上執(zhí)行:
yarn-daemon.sh
start
nodemanagerBONCInformation
Change
t2h8eWorld單服務(wù)停止命令(一)啟動zookeeper服務(wù)所有zookeeper服務(wù)節(jié)點執(zhí)行:zkServer.sh
stop啟動JournalNode服務(wù)所有journalenode節(jié)點執(zhí)行:hadoop-daemon.sh
stop
journalnode啟動NameNode服務(wù)所有NameNode節(jié)點執(zhí)行hadoop-daemon.sh
stop
namenode啟動DataNode服務(wù)在某一個節(jié)點上執(zhí)行:hadoop-daemons.sh
stop
datanode或者在所有DataNode節(jié)點上執(zhí)行:(某些時候需要單獨關(guān)閉某個節(jié)點的服務(wù))
hadoop-daemon.sh
stop
datanodeBONCInformation
Change
t2h9eWorld單服務(wù)停止命令(二)停止ZooKeeperFailoverController在所有NameNode節(jié)點上執(zhí)行:
hadoop-daemon.sh
stop
zkfc停止ResourceManager在所有resoucermanager節(jié)點上執(zhí)行:
yarn-daemon.sh
stop
resoucermanager停止nodemanager在所有nodemanager節(jié)點上執(zhí)行:yarn-daemon.sh
stop
nodemanagerBONC啟停hadoop集群在NameNode節(jié)點創(chuàng)建
啟動、停止[hadoop@hadoop001
~]$
mkdir
/opt/beh/host_dir[hadoop@hadoop001
~]$cd
/opt/beh/host_dirInformation
Change
t3h0eWorldBONCInformation
Change
t3h1eWorld配置各服務(wù)HOST文件[hadoop@hadoop001
host_dir]$
cat
host_zookeeperhadoop001hadoop002hadoop003[hadoop@hadoop001
host_dir]$
cat
host_journalnodehadoop001hadoop002hadoop003[hadoop@hadoop001
host_dir]$
cat
host_namenodehadoop001hadoop002[hadoop@hadoop001
host_dir]$
cat
host_datanodehadoop003hadoop004hadoop005hadoop006hadoop007hadoop008[hadoop@hadoop001
host_dir]$
cat
host_resourcemanagerhadoop001hadoop002[hadoop@hadoop001
host_dir]$
cat
host_nodemanagerhadoop003hadoop004hadoop005hadoop006hadoop007hadoop008BONC集群
啟動[hadoop@bonc1
host_dir]$
sh
start-all.shJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStarting
zookeeper
...STARTEDJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStarting
zookeeper
...STARTEDJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStarting
zookeeper
...STARTEDstarting
journalnode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-journalnode-bonc1.outstarting
journalnode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-journalnode-bonc2.outstarting
journalnode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-journalnode-bonc3.outstarting
namenode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-namenode-bonc1.outstarting
zkfc,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-zkfc-bonc1.outstarting
namenode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-namenode-bonc2.outstarting
zkfc,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-zkfc-bonc2.outbonc3:
starting
datanode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-datanode-bonc3.outbonc5:
starting
datanode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-datanode-bonc5.outbonc4:
starting
datanode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-datanode-bonc4.outbonc6:
starting
datanode,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-datanode-bonc6.outsafe_mode_is_offstarting
jobtrackerha,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-jobtrackerha-bonc1.outstarting
mrzkfc,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-mrzkfc-bonc1.outstarting
jobtrackerha,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-jobtrackerha-bonc2.outstarting
mrzkfc,
logging
to
/opt/beh/hadoop/logs/hadoop-hadoop-mrzkfc-bonc2.outbonc5:
starting
tasktracker,
logging
to
/opt/beh/hadoop/bin-mapreduce1/../logs/hadoop-hadoop-tasktracker-bonc5.outbonc4:
starting
tasktracker,
logging
to
/opt/beh/hadoop/bin-mapreduce1/../logs/hadoop-hadoop-tasktracker-bonc4.outbonc3:
starting
tasktracker,
logging
to
/opt/beh/hadoop/bin-mapreduce1/../logs/hadoop-hadoop-tasktracker-bonc3.outbonc6:
starting
tasktracker,
logging
to
/opt/beh/hadoop/bin-mapreduce1/../logs/hadoop-hadoop-tasktracker-bonc6.outInformation
Change
t3h2eWorldBONC集群
停止[hadoop@bonc1
host_dir]$
sh
top_all.shbonc4:
stopbonc5:
stopbonc3:
stopbonc6:
stoptasktrackertasktrackertasktrackertasktrackerstopstopstopstopmrzkfcjobtrackerhamrzkfcjobtrackerhabonc5:
stopbonc3:
stopbonc4:
stopbonc6:
stopdatanodedatanodedatanodedatanodenamenodezkfcnamenodezkfcstopstopstopstopstopstopstopjournalnodejournalnodejournalnodeJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStop zookeeper
...STOPPEDJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStop zookeeper
...STOPPEDJMX
enabled
by
defaultUsing
config:
/opt/beh/zookeeper/bin/../conf/zoo.cfgStop zookeeper
...STOPPEDInformation
Change
t3h3eWorldBONCNameNodeweb端口為50070,登陸http://{NameNode-ip-address}:50070可以查看NameNode相關(guān)信息,包括是否為活躍節(jié)點,查看當(dāng)前hdfs占用的空間占比,以及NameNode節(jié)點各種日志.Information
Change
t3h4eWorldBONChDfs空間使用情況Information
Change
t3h5eWorldBONCNameNode節(jié)點日志Information
Change
t3h6eWorldoBONCJob
(一)web端口為19888,登陸http://{NameNode-ip-address}:19888可以查看正在執(zhí)行、結(jié)束、失敗的job以及job的各種詳細(xì)信息,包括執(zhí)行job時的各種配置參數(shù)、map數(shù)、reduce數(shù)、 文件大小、寫入文件大小等等.針對失敗的job進(jìn)行日志以及參數(shù)等分析,查明原因作出相應(yīng)對策.rldBONCHadoop集群常用命令(一)hdfsdfsadmin查看集群各節(jié)點狀態(tài)[hadoop@hadoop009
~]$
hdfs
dfsadmin-reportConfigured
Capacity:
63749735645184
(57.98TB)Present
Capacity:
60496306655232
(55.02
TB)DFS
Remaining:
36531939196928
(33.23
TB)DFS
Used:
23964367458304
(21.80
TB)DFS
Used%:
39.61%Under
replicated
blocks:
275Blocks
with
corrupt
replicas:
0Missing
blocks:
0Datanodes
available:
6
(6
total,
0
dead)Live
datanodes:Name:
66:50010
(hadoop008)Hostname:
hadoop008mission
Status
:
NormalConfigured
Capacity:
10624955940864
(9.66TB)DFS
Used:
3972594683904
(3.61
TB)Non
DFS
Used:
542238167040
(505.00
GB)DFS
Remaining:
6110123089920
(5.56
TB)DFS
Used%:
37.39%DFS
Remaining%:
57.51%Configured
Cache
Capacity:
0
(0
B)CacheUsed:
0
(0
B)Cache
Remaining:
0
(0
B)Cache
Used%:
100.00%CacheRemaining%:
0.00%Last
contact:
Mon
May
19
01:01:42
CST
2014Information
Change
t3h8eWorldBONCInformation
Change
t3h9eWorldHadoop集群常用命令(二)查看安全模式hdfs
dfsadmin
-safemode
enter
|
leave
|
get
|
wait[hadoop@hadoop009
~]$hdfs
dfsadmin
-safemode
getSafe
modeisOFF查看namenode是否活躍[hadoop@hadoop009
~]$hdfshaadmin
-getServiceState
nn1active[hadoop@hadoop009
~]$hdfshaadmin
-getServiceState
nn2Standby查看hdfs的參數(shù)配置[hadoop@hadoop009~]$
hdfs
getconf-confKeydfs.replication3BONCInformation
Change
t4h0eWorldHadoop集群常用命令(三)hdfs
fsck檢測文件系統(tǒng)的塊信息[hadoop@hadoop009
~]$
hdfs
fsck
/...Status:
HEALTHYTotal
size:
7917560996782
BTotal
dirs:
1047Total
files:
36203Total
symlinks:
0Total
blocks
(validated): 61095
(avg.
block
size
129594254
B)Minimally
replicated
blocks: 61095
(100.0
%)Over-replicated
blocks:Under-replicated
blocks:0
(0.0
%)275
(0.45011866
%)Mis-replicatedblocks: 0
(0.0
%)Default
replicationfactor:
3Average
block
replication:
3.01350360Corrupt
blocks:
Missing
replicas:Number
of
data-nodes:1100
(0.5939204
%)6Number
of
racks:
1FSCK
ended
at
Mon
May
19
01:13:10
CST
2014
in
889
millisecondsBONCInformation
Change
t4h1eWorldHadoop集群常用命令(四)查看jobtracker節(jié)點的活躍狀態(tài)[hadoop@hadoop009
~]$
hadoop
mrhaadmin-getServiceState
jt1active[hadoop@hadoop009
~]$
hadoop
mrhaadmin-getServiceState
jt2Standby查看執(zhí)行job的信息
[hadoop@hadoop009
~]$
mapred
job-list1
jobs
currently
runningJobId
State
StartTime
UserName
Priority
SchedulingInfojob_201405181945_0002
4
1400433843326 hadoop
NORMAL
NA[hadoop@hadoop009
~]$
mapred
job
-list1
jobs
currently
runningPriority
SchedulingInfoJobId
State
StartTime
UserNamejob_201405181945_0002
11400433843326 hadoop
NORMAL
NABONCInformation
Change
t4h2eWorldHadoop集群常用命令(五)殺死job[hadoop@hadoop009
~]$
mapred
job
-list1
jobs
currently
runningJobId
State
StartTime
UserNamePrioritySchedulingInfojob_201405181945_0004
4
1400434102248 hadoop
NORMAL
NA[hadoop@hadoop009
~]$
mapred
job
-kill
job_201405181945_0004Killed
jobjob_201405181945_0004[hadoop@hadoop009
~]$mapred
job
-list0
jobs
currently
runningJobId
State
StartTime
UserName
Priority
SchedulingInfoBONCInformation
Change
t4h3eWorldHadoop集群常用命令(六)修改job的執(zhí)行優(yōu)先級[hadoop@hadoop009
~]$
hadoop
job
-list1
jobs
currently
runningJobId
State
StartTime
UserNamePrioritySchedulingInfojob_201405181945_0005
4
1400434312199 hadoop
NORMAL
NA[hadoop@hadoop009
~]$hadoop
job
-set-priorityjob_201405181945_0005VERY_HIGHChanged
jobpriority.[hadoop@hadoop009
~]$
hadoop
job
-list1
jobs
currently
runningPriority
SchedulingInfoJobId
State
StartTime
UserNamejob_201405181945_0005
11400434312199 hadoop
VERY_HIGH
NABONCInformation
Change
t4h4eWorldHadoop集群常用命令(七)集群數(shù)據(jù)節(jié)點平衡開啟、關(guān)閉命令[hadoop@hadoop009
~]$
start-balancer.sh
或者
hdfs
balancestartingbalancer,
loggingto
/opt/beh/hadoop/logs/hadoop-hadoop-balancer-hadoop009.out[hadoop@hadoop009
~]$
more
/opt/beh/hadoop/logs/hadoop-hadoop-balancer-hadoop009.outTime
Stamp Iteration#
Bytes
Already
Moved
Bytes
Left
To
Move
BytesBeing
MovedThe
cluster
is
balanced.
Exiting...Balancing
took
1.902
seconds(blocks,
-f)
unlimited(-i)
773705pending
signals
max
locked
memorymax
memorysize(kbytes,
-l)
64(kbytes,
-m)
unlimitedopen
files(-n)
131072pipe
size (512
bytes,
-p)
8POSIX
message
queues (bytes,
-q)
819200real-time
prioritystack
size(-r)
0(kbytes,
-s)
10240cpu
time (seconds,
-t)
unlimited(-u)
unlimitedmax
user
processesvirtual
memoryfile
locks(kbytes,
-v)
unlimited(-x)
unlimitedBONCHadoop集群常用命令(八)檢測所有節(jié)點已經(jīng)開啟的服務(wù)[hadoop@hadoop009
~]$
sh
all_jps.sh--------hadoop001----------QuorumPeerMainJournalNodeNameNodeDFSZKFailoverControllerJobTrackerHADaemonMRZKFailoverController--------hadoop002----------QuorumPeerMainJournalNodeNameNodeDFSZKFailoverControllerJobTrackerHADaemonMRZKFailoverController--------hadoop003----------QuorumPeerMainJournalNodeDataNodeTaskTracker--------hadoop004----------DataNodeTaskTracker--------hadoop005----------DataNodeTaskTracker--------hadoop006----------DataNodeTaskTracker--------hadoop007----------DataNodeTaskTracker--------hadoop008----------DataNodeTaskTracker--------hadoop009----------RunJarRunJarRunJarBootstrap--------hadoop010----------RunJarBootstrapInformation
Change
t4h5eWorldBONCNameNode節(jié)點日志相關(guān)[hadoop@hadoop001logs]$
cd
/opt/beh/logs/hadoop[hadoop@hadoop001
logs]$
pwd/opt/beh/logs/hadoop[hadoop@hadoop001
logs]$
ll##namenode進(jìn)程對應(yīng)日志-rw-rw-r--
1
hadoop
hadoop
188440522
Sep
12
12:51
hadoop-hadoop-namenode-hadoop001.logInformation
Change
the
WorldBONCInformation
Change
the
WorldDataNode節(jié)點日志相關(guān)[hadoop@hadoop001
logs]$
cd
/opt/beh/logs/hadoop[hadoop@hadoop001
logs]$
pwd/opt/beh/logs/hadoop[hadoop@hadoop001
logs]$
ls-lrt
|grep
log
|grep-vlog.##DataNode進(jìn)程對應(yīng)日志-rw-r--r-- 1
hadoop
hadoop 259748137
May
20
17:19
hadoop-hadoop-tasktracker-hadoop004.log##TaskTracker進(jìn)程對應(yīng)日志-rw-r--r-- 1
hadoop
hadoop 132643847
May
20
17:19
hadoop-hadoop-datanode-hadoop004.logBONC集群負(fù)載
(一)主要CPU以及l(fā)oad負(fù)載.合理的調(diào)整ETL流程能有效的較少Job集中對整個流程加工的時間延遲.Load值一CPU消耗過高達(dá)到95%以上時,不宜增加新的Job。一般Load值低于CPU邏輯線程數(shù)的3倍,超過
3倍時服務(wù)器間通信會變得較慢。Information
Change
t4
溫馨提示
- 1. 本站所有資源如無特殊說明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請下載最新的WinRAR軟件解壓。
- 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶所有。
- 3. 本站RAR壓縮包中若帶圖紙,網(wǎng)頁內(nèi)容里面會有圖紙預(yù)覽,若沒有圖紙預(yù)覽就沒有圖紙。
- 4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
- 5. 人人文庫網(wǎng)僅提供信息存儲空間,僅對用戶上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對用戶上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對任何下載內(nèi)容負(fù)責(zé)。
- 6. 下載文件中如有侵權(quán)或不適當(dāng)內(nèi)容,請與我們聯(lián)系,我們立即糾正。
- 7. 本站不保證下載資源的準(zhǔn)確性、安全性和完整性, 同時也不承擔(dān)用戶因使用這些下載資源對自己和他人造成任何形式的傷害或損失。
最新文檔
- 人教部編版六年級語文上冊第18課《古詩三首》精美課件
- 2024年石家莊駕駛員客運從業(yè)資格證模擬考試題
- 2024年雙鴨山從業(yè)資格證模擬考試題庫
- 2024年臨夏駕??荚嚳瓦\從業(yè)資格證考試題庫
- 浙江省七彩陽光聯(lián)盟2022-2023學(xué)年高二11月期中生物試題(原卷版)
- 2024年工廠研發(fā)與技術(shù)轉(zhuǎn)讓合同
- 計算機等級考試一級計算機基礎(chǔ)及WPS Office應(yīng)用試題與參考答案(2025年)
- 三年級上冊數(shù)學(xué)四邊形的認(rèn)識說課稿范文
- 京劇批評研究的回顧、新變與不足
- 感謝信,給我?guī)砉ぷ鳈C會
- 中國 美國 日本水洗標(biāo)志對比
- 新產(chǎn)品試制流程管理辦法
- 通用橫版企業(yè)報價單模板
- 潛油泵及潛油泵加油機講義
- 物業(yè)服務(wù)公司各崗位規(guī)范用語
- 醫(yī)患溝通內(nèi)容要求記錄模板(入院、入院三日、術(shù)前、術(shù)后、出院)
- 航海學(xué)天文定位第四篇第6章天文定位
- 淺談深度教學(xué)中小學(xué)數(shù)學(xué)U型學(xué)習(xí)模式
- 物理電學(xué)暗箱專題30道
- 裝修公司員工勞動合同
- 江西上饒鉛山汽車駕駛科目三考試線路
評論
0/150
提交評論