1. 程式人生 > 實用技巧 >hadoop 單臺機器偽叢集

hadoop 單臺機器偽叢集

  1. 網路設定

  1. /etc/sysconfig/network-scripts/ 目錄下

    ifcfg-eth0檔案

    BOOTPROTO="static"
    DEFROUTE="yes"
    IPADDR=192.168.1.10
    GATEWAY=192.168.1.1
    DNS1=192.168.1.1
    NETMASK=255.255.255.0
    ONBOOT="yes"
    
  2. /etc/ 目錄下

    hosts檔案

    192.168.1.10 hadoop10
    
  3. systemctl restart network

  4. 設定環境變數

    vim /etc/profile

    1. java

      ##java_home
      JAVA_HOME=/opt/module/jdk1.8.0_151
      JRE_HOME=$JAVA_HOME/jre
      CLASS_PATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar:$JRE_HOME/lib
      PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
      export JAVA_HOME JRE_HOME CLASS_PATH PATH
      

      source /etc/profile

      java -version 測試

    2. hadoop

      ##hadoop_home
      export HADOOP_HOME=/opt/module/hadoop-2.10.1
      export PATH=$PATH:$HADOOP_HOME/bin
      export PATH=$PATH:$HADOOP_HOME/sbin
      

      hadoop 測試

  5. 修改配置檔案

    1. hadoop/etc/hadoop/core-site.xml

      <configuration>
          <property>
        <!--指定namenode的地址-->
      	    <name>fs.defaultFS</name>
          	<value>hdfs://127.0.0.1:9000</value>
          </property>
      	<property>
        <!--指定namenode的地址-->
      	    <name>hadoop.tmp.dir</name>
          	<value>/opt/module/hadoop-2.10.1/data/temp</value>
          </property>
      	<!-- 當前使用者全設定成root -->	
          <property>
              <name>hadoop.http.staticuser.user</name>
              <value>root</value>
          </property>
      	<!-- 不開啟許可權檢查 -->
          <property>
              <name>dfs.permissions.enabled</name>
              <value>false</value>
          </property>
      </configuration>
      
    2. hadoop/etc/hadoop/hdfs-site.xml

      	<!-- 指定hdfs副本的數量 -->
      	<property>
              <name>dfs.replication</name>
              <value>1</value>
          </property>
      
    3. hadoop/etc/hadoop/yarn-site.xml

      <configuration>
          <property>
              <name>yarn.nodemanager.aux-services</name>
              <value>mapreduce_shuffle</value>
          </property>
          <property>
              <name>yarn.resourcemanager.hostname</name>
              <value>hadoop10</value>
          </property>
      
          <property>
              <name>yarn.log-aggregation-enable</name>
              <value>true</value>
          </property>
          <property>
              <name>yarn.log-aggregation-retain-seconds</name>
              <value>604800</value>
          </property>
      </configuration>
      
    4. hadoop/etc/hadoop/mapred-site.xml

      <configuration>
          <property>
              <name>mapreduce.framework.name</name>
              <value>yarn</value>
          </property>
          <property>
              <name>mapreduce.jobhistory.address</name>
              <value>hadoop10:10020</value>
          </property>
          <property>
              <name>mapreduce.jobhistory.webapp.address</name>
              <value>hadoop10:19888</value>
          </property>
      </configuration>
      
  6. 啟動

    hdfs namenode -format
    sbin/hadoop-daemon.sh start namenode
    ./hadoop-daemon.sh start datanode
    
    http://192.168.1.10:50070 訪問測試
    
    sbin/yarn-daemon.sh start resourcemanager
    sbin/yarn-daemon.sh start nodemanager
    
    sbin/mr-jobhistory-daemon.sh start historyserver
    
    hdfs dfs -put /opt/module/hadoop-2.10.1/share/test.txt /input
    將linux目錄中的檔案,上傳到hdfs中
    
    hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.10.1.jar wordcount /input /output
    
  7. 檢視

    http://192.168.1.10:50070 hdfs

    http://192.168.1.10:8088 hadoop叢集

    http://hadoop10:19888 歷史