1. 程式人生 > >Hadoop安裝之standAlone單機

Hadoop安裝之standAlone單機

hadoop安裝1.x和2.x有三種安裝架構,本文將介紹第一種安裝模式standAlone

一:standAlone(單機)

standAlone安裝將所有服務都安裝在一臺機器上,如下:

執行服務

伺服器IP

NameNode

192.168.254.100

SecondaryNameNode

192.168.254.100

DataNode

192.168.254.100

ResourceManager

192.168.254.100

NodeManager

192.168.254.100

1)下載安裝:

JDK的安裝請自行安裝。

下載地址:http://archive.apache.org/dist/hadoop/common/hadoop-2.7.5/hadoop-2.7.5.tar.gz

解壓:

mkdir -p /export/softwares

mkdir -p /exprot/servers

cd /export/softwares

tar -zxvf hadoop-2.7.5.tar.gz -C ../servers/

2)修改配置檔案:

2.1:修改core-site.xml

cd /export/servers/hadoop-2.7.5/etc/hadoop

vim core-site.xml

<configuration>

<property>

<name>fs.default.name</name>

<!--  namenode所在機器地址 -->

<value>hdfs://192.168.254.100:8020</value>

</property>

<property>

<name>hadoop.tmp.dir</name>

<value>/export/servers/hadoop-2.7.5/hadoopDatas/tempDatas</value>

</property>

<!--  緩衝區大小,可根據伺服器效能動態調整 -->

<property>

<name>io.file.buffer.size</name>

<value>4096</value>

</property>

 

<!--  開啟hdfs的垃圾桶機制,刪除掉的資料可以從垃圾桶中回收,單位分鐘 -->

<property>

<name>fs.trash.interval</name>

<value>10080</value>

</property>

</configuration>

2.2:修改hdfs-site.xml

cd /export/servers/hadoop-2.7.5/etc/hadoop

vim hdfs-site.xml

<configuration>

<!-- NameNode儲存元資料資訊的路徑,一般先確定磁碟的掛載目錄,然後多個目錄用,進行分割   -->

<!--   叢集動態上下線

<property>

<name>dfs.hosts</name>

<value>/export/servers/hadoop-2.7.4/etc/hadoop/accept_host</value>

</property>

<property>

<name>dfs.hosts.exclude</name>

<value>/export/servers/hadoop-2.7.4/etc/hadoop/deny_host</value>

</property>

 -->

 

 <property>

<name>dfs.namenode.secondary.http-address</name>

<value>node01:50090</value>

</property>

 

<property>

<!--瀏覽器訪問hdfs埠-->

<name>dfs.namenode.http-address</name>

<value>node01:50070</value>

</property>

<property>

<name>dfs.namenode.name.dir</name>

<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas,file:///export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas2</value>

</property>

<!--  定義dataNode資料儲存的節點位置,實際工作中,一般先確定磁碟的掛載目錄,然後多個目錄用,進行分割  -->

<property>

<name>dfs.datanode.data.dir</name>

<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas,file:///export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas2</value>

</property>

<property>

<name>dfs.namenode.edits.dir</name>

<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/nn/edits</value>

</property>

 

<property>

<name>dfs.namenode.checkpoint.dir</name>

<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/snn/name</value>

</property>

<property>

<name>dfs.namenode.checkpoint.edits.dir</name>

<value>file:///export/servers/hadoop-2.7.5/hadoopDatas/dfs/snn/edits</value>

</property>

<!--副本數量-->

<property>

<name>dfs.replication</name>

<value>3</value>

</property>

<!--hadoop許可權校驗-->

<property>

<name>dfs.permissions</name>

<value>false</value>

</property>

<!--block塊大小,128M-->

<property>

<name>dfs.blocksize</name>

<value>134217728</value>

</property>

</configuration>

2.3:修改hadoop-env.sh

cd /exprot/servers/hadoop2.7.5/etc/hadoop

vim hadoop-evn.sh

export JAVA_HOME=/export/servers/jdk1.8.0_141

2.4:修改mapredp-site.xml

cd /exprot/servers/hadoop2.7.5/etc/hadoop

vim mapredp-site.xml

<configuration>

<property>

<name>mapreduce.framework.name</name>

<value>yarn</value>

</property>

 

<property>

<name>mapreduce.job.ubertask.enable</name>

<value>true</value>

</property>

<property>

<name>mapreduce.jobhistory.address</name>

<value>192.168.254.100:10020</value>

</property>

 

<property>

<name>mapreduce.jobhistory.webapp.address</name>

<value>192.168.254.100:19888</value>

</property>

</configuration>

2.5:修改yarn-site.xml

cd /exprot/servers/hadoop2.7.5/etc/hadoop

vim yarn-site.xml

<configuration>

<property>

<name>yarn.resourcemanager.hostname</name>

<value>node01</value>

</property>

<property>

<name>yarn.nodemanager.aux-services</name>

<value>mapreduce_shuffle</value>

</property>

<property>

<name>yarn.log-aggregation-enable</name>

<value>true</value>

</property>

<property>

<name>yarn.log-aggregation.retain-seconds</name>

<value>604800</value>

</property>

</configuration>

2.6:修改mapred-env.sh

cd /exprot/servers/hadoop2.7.5/etc/hadoop

vim mapred-env.sh

export JAVA_HOME=/export/servers/jdk1.8.0_141

2.7:slaves

cd /exprot/servers/hadoop2.7.5/etc/hadoop

vim slaves

localhost

3)啟動叢集

啟動hadoop叢集,需要啟動hdfs和yarn兩個模組。注意首次啟動HDFS時,必須對其進行格式化操作,本質上是一些清理和準備工作,因為此時HDFS在物理上還是不存在的

格式化命令:hdfs namenode -format

再啟動前需要建立存放資料資料夾

cd  /export/servers/hadoop-2.7.5

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/tempDatas

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/namenodeDatas2

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/datanodeDatas2

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/nn/edits

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/snn/name

mkdir -p /export/servers/hadoop-2.7.5/hadoopDatas/dfs/snn/edits

啟動命令:

cd  /export/servers/hadoop-2.7.5/

bin/hdfs namenode -format --如果已經格式化了就不需要再格式化

sbin/start-dfs.sh

sbin/start-yarn.sh

sbin/mr-jobhistory-daemon.sh start historyserver

4)介面檢視:

http://192.168.254.100:50070/explorer.html#/  檢視hdfs

http://192.168.254.100:8088/cluster   檢視yarn叢集

http://192.168.254.100:19888/jobhistory  檢視歷史完成的任務