Hadoop安裝之standAlone單機
hadoop安裝1.x和2.x有三種安裝架構,本文將介紹第一種安裝模式standAlone
一:standAlone(單機)
standAlone安裝將所有服務都安裝在一臺機器上,如下:
執行服務 |
伺服器IP |
NameNode |
192.168.254.100 |
SecondaryNameNode |
192.168.254.100 |
DataNode |
192.168.254.100 |
ResourceManager |
192.168.254.100 |
NodeManager |
192.168.254.100 |
1)下載安裝:
JDK的安裝請自行安裝。
下載地址:http://archive.apache.org/dist/hadoop/common/hadoop-2.7.5/hadoop-2.7.5.tar.gz
解壓:
mkdir -p /export/softwares
mkdir -p /exprot/servers
cd /export/softwares
tar -zxvf hadoop-2.7.5.tar.gz -C ../servers/
2)修改配置檔案:
2.1:修改core-site.xml
cd /export/servers/hadoop-2.7.5/etc/hadoop
vim core-site.xml
<configuration>
<property>
<name>fs.default.name</name>
<!-- namenode所在機器地址 -->
<value>hdfs://192.168.254.100:8020</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/export/servers/hadoop-2.7.5/hadoopDatas/tempDatas</value>
</property>
<!-- 緩衝區大小,可根據伺服器效能動態調整 -->
<property>
<name>io.file.buffer.size</name>
<value>4096</value>
</property>
<!-- 開啟hdfs的垃圾桶機制,刪除掉的資料可以從垃圾桶中回收,單位分鐘 -->
<property>
<name>fs.trash.interval</name>
<value>10080</value>
</property>
</configuration>
2.2:修改hdfs-site.xml
cd /export/servers/hadoop-2.7.5/etc/hadoop
vim hdfs-site.xml
<configuration>
<!-- NameNode儲存元資料資訊的路徑,一般先確定磁碟的掛載目錄,然後多個目錄用,進行分割 -->
<!-- 叢集動態上下線
<property>
<name>dfs.hosts</name>
<value>/export/servers/hadoop-2.7.4/etc/hadoop/accept_host</value>
</property>
<property>
<name>dfs.hosts.exclude</name>
<value>/export/servers/hadoop-2.7.4/etc/hadoop/deny_host</value>
</property>
-->
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>node01:50090</value>
</property>
<property>
<!--瀏覽器訪問hdfs埠-->
<name>dfs.namenode.http-address</name>
<value>node01:50070</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas,file:///export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas2</value>
</property>
<!-- 定義dataNode資料儲存的節點位置,實際工作中,一般先確定磁碟的掛載目錄,然後多個目錄用,進行分割 -->
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas,file:///export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas2</value>
</property>
<property>
<name>dfs.namenode.edits.dir</name>
<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/nn/edits</value>
</property>
<property>
<name>dfs.namenode.checkpoint.dir</name>
<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/snn/name</value>
</property>
<property>
<name>dfs.namenode.checkpoint.edits.dir</name>
<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/dfs/snn/edits</value>
</property>
<!--副本數量-->
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<!--hadoop許可權校驗-->
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>
<!--block塊大小,128M-->
<property>
<name>dfs.blocksize</name>
<value>134217728</value>
</property>
</configuration>
2.3:修改hadoop-env.sh
cd /exprot/servers/hadoop2.7.5/etc/hadoop
vim hadoop-evn.sh
export JAVA_HOME=/export/servers/jdk1.8.0_141
2.4:修改mapredp-site.xml
cd /exprot/servers/hadoop2.7.5/etc/hadoop
vim mapredp-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.job.ubertask.enable</name>
<value>true</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>192.168.254.100:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>192.168.254.100:19888</value>
</property>
</configuration>
2.5:修改yarn-site.xml
cd /exprot/servers/hadoop2.7.5/etc/hadoop
vim yarn-site.xml
<configuration>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>node01</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.log-aggregation-enable</name>
<value>true</value>
</property>
<property>
<name>yarn.log-aggregation.retain-seconds</name>
<value>604800</value>
</property>
</configuration>
2.6:修改mapred-env.sh
cd /exprot/servers/hadoop2.7.5/etc/hadoop
vim mapred-env.sh
export JAVA_HOME=/export/servers/jdk1.8.0_141
2.7:slaves
cd /exprot/servers/hadoop2.7.5/etc/hadoop
vim slaves
localhost
3)啟動叢集
啟動hadoop叢集,需要啟動hdfs和yarn兩個模組。注意首次啟動HDFS時,必須對其進行格式化操作,本質上是一些清理和準備工作,因為此時HDFS在物理上還是不存在的
格式化命令:hdfs namenode -format
再啟動前需要建立存放資料資料夾
cd /export/servers/hadoop-2.7.5
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/tempDatas
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas2
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas2
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/nn/edits
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/snn/name
mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/dfs/snn/edits
啟動命令:
cd /export/servers/hadoop-2.7.5/
bin/hdfs namenode -format --如果已經格式化了就不需要再格式化
sbin/start-dfs.sh
sbin/start-yarn.sh
sbin/mr-jobhistory-daemon.sh start historyserver
4)介面檢視:
http://192.168.254.100:50070/explorer.html#/ 檢視hdfs
http://192.168.254.100:8088/cluster 檢視yarn叢集
http://192.168.254.100:19888/jobhistory 檢視歷史完成的任務