1. 程式人生 > >hadoop錯誤org.apache.hadoop.yarn.exceptions.YarnException Unauthorized request to start container

hadoop錯誤org.apache.hadoop.yarn.exceptions.YarnException Unauthorized request to start container

錯誤:

17/11/22 15:17:15 INFO client.RMProxy: Connecting to ResourceManager at Master/192.168.136.100:8032
17/11/22 15:17:16 INFO input.FileInputFormat: Total input paths to process : 1
17/11/22 15:17:16 WARN hdfs.DFSClient: Caught exception 
java.lang.InterruptedException
        at java.lang.Object.wait(Native Method)
        at java.lang.Thread.join(Thread.java:1252)
        at java.lang.Thread.join(Thread.java:1326)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.closeResponder(DFSOutputStream.java:609)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.endBlock(DFSOutputStream.java:370)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:546)
17/11/22 15:17:17 INFO mapreduce.JobSubmitter: number of splits:1
17/11/22 15:17:17 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1511275253885_0004
17/11/22 15:17:17 INFO impl.YarnClientImpl: Submitted application application_1511275253885_0004
17/11/22 15:17:17 INFO mapreduce.Job: The url to track the job: http://Master:8088/proxy/application_1511275253885_0004/
17/11/22 15:17:17 INFO mapreduce.Job: Running job: job_1511275253885_0004
17/11/22 15:17:19 INFO mapreduce.Job: Job job_1511275253885_0004 running in uber mode : false
17/11/22 15:17:19 INFO mapreduce.Job:  map 0% reduce 0%
17/11/22 15:17:19 INFO mapreduce.Job: Job job_1511275253885_0004 failed with state FAILED due to: Application application_1511275253885_0004 failed 2 times due to Error launching appattempt_1511275253885_0004_000002. Got exception: org.apache.hadoop.yarn.exceptions.YarnException: Unauthorized request to start container. 
This token is expired. current time is 1511363838355 found 1511335638708
Note: System times on machines may be out of sync. Check system time and time zones.
        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
        at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
        at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
        at org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:168)
        at org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106)
        at org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:122)
        at org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:250)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
. Failing the application.
17/11/22 15:17:19 INFO mapreduce.Job: Counters: 0
[
[email protected]
mapreduce]# cp /usr/share/zoneinfo/Asia/Shanghai /etc/localtime cp: "/usr/share/zoneinfo/Asia/Shanghai" 與"/etc/localtime" 為同一檔案 [[email protected] mapreduce]# ntpdate pool.ntp.org 22 Nov 15:21:30 ntpdate[28360]: step time server 193.228.143.22 offset -11.343113 sec [[email protected] mapreduce]# ^C [
[email protected]
mapreduce]# ntpdate pool.ntp.org 22 Nov 15:22:12 ntpdate[28365]: adjust time server 193.228.143.22 offset 0.004859 sec [[email protected] mapreduce]# ntpdate s2c.time.edu.cn 22 Nov 15:31:59 ntpdate[28423]: adjust time server 202.112.10.36 offset -0.017539 sec
原因:

    namenode,datanode時間同步問題

解決辦法

    多個datanode與namenode進行時間同步,在每臺伺服器(master slave1 slaves2)執行如下兩個命令進行同步

   例如:

[[email protected] mapreduce]# cp /usr/share/zoneinfo/Asia/Shanghai /etc/localtime
cp:是否覆蓋"/etc/localtime"? y
[[email protected] mapreduce]# ntpdate s2c.time.edu.cn
注意:如果提示說:
[[email protected] mapreduce]# ntpdate pool.ntp.org                               
-bash: ntpdate: command not found

應該先下載ntpdate:

yum install -y ntp



相關推薦

hadoop錯誤org.apache.hadoop.yarn.exceptions.YarnException Unauthorized request to start container

錯誤: 17/11/22 15:17:15 INFO client.RMProxy: Connecting to ResourceManager at Master/192.168.136.100:8032 17/11/22 15:17:16 INFO input.Fil

hadoop上傳檔案錯誤org.apache.hadoop.ipc.RemoteException(java.io.IOException)

部落格引用處(以下內容在原有部落格基礎上進行補充或更改,謝謝這些大牛的部落格指導): hadoop上傳檔案錯誤org.apache.hadoop.ipc.RemoteException(java.io.IOException) 搭建好hadoop後使用hadoop fs -put 命令上

Hbase 出現 org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet 錯誤

hadoop nbsp mode sta oop proto 出現 method cep ERROR: org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet

執行HBase shell時出現ERROR: org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet錯誤解決辦法(圖文詳解)

cep ESS 關註 align comm util code ade dap   不多說,直接上幹貨! [kfk@bigdata-pro01 bin]$ jps 1968 NameNode 2385 ResourceManager 2259 Jou

hive中刪除表的錯誤Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException

成了 ret jar tor java-5 drop meta org -c hive使用drop table 表名刪除表時報錯,return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException

eclipse執行mapereduce程式時報如下錯誤org.apache.hadoop.io.nativeio.NativeIO$Windows.createDirectoryWithMode0(

eclipse執行mapereduce程式時報如下錯誤: log4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN

Hive建立表格報Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException的錯誤

1.首先保證hive正常配置, 修改$HIVE_HOME/conf/hive-site.xml <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?

hadoop寫MR程式報錯java.lang.AbstractMethodError: org.apache.hadoop.yarn.api.records.LocalResource.setShou

情況:在本地書寫mapreduce的時候,執行driver類 開始跑任務的時候,有時候可能會報 java.lang.AbstractMethodError: org.apache.hadoop.yarn.api.records.LocalResource.setShouldBeUploadedT

hive安裝完MySQL後報Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient錯誤

錯誤提示: Exception in thread "main" java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.

hadoop解決Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/hadoop/yarn/util/Apps

linux+eclipse+本地執行WordCount丟擲下面異常: Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/hadoop/yarn/util/Apps。 解決:沒有把yar

hadoop3.1.1下MapReduce操作出現錯誤: 找不到或無法載入主類org.apache.hadoop.mapreduce.v2.app.MRAppMaster 問題解決方法

Hadoop3.1.1下成功安裝後,進行MapReduce操作出現錯誤: 錯誤: 找不到或無法載入主類org.apache.hadoop.mapreduce.v2.app.MRAppMaster 解決辦法: 在命令列下輸入如下命令,並將返回的地址複製。 hadoop c

錯誤: 找不到或無法載入主類 org.apache.hadoop.hdfs.server.namenode.NameNode 問題解決

問題描述:   執行指令 在hadoop安裝路徑下執行  bin/hdfs namenode -format 時,出現“錯誤:找不到或無法載入主類org.apache.hadoop.hdfs.server.namenode.NameNode" 問題分析:   此問題是由於在

HBASE啟動失敗,Failed construction of Master: class org.apache.hadoop.hbase.master.HMaster

sta caused and tex runtime sla class -1 regions Master日誌錯誤:2015-12-02 06:34:32,394 ERROR [main] master.HMasterCommandLine: Master exiting

解決kylin報錯 ClassCastException org.apache.hadoop.hive.ql.exec.ConditionalTask cannot be cast to org.apache.hadoop.hive.ql.exec.mr.MapRedTask

conf lan exe hive oop ann 關於 .exe map 方法:去掉參數SET hive.auto.convert.join=true; 從配置文件$KYLIN_HOME/conf/kylin_hive_conf.xml刪掉 或 kylin-gui的cu

排查Hive報錯:org.apache.hadoop.hive.serde2.SerDeException: java.io.IOException: Start of Array expected

arr .json span 問題 catalog pan 不支持 led open CREATE TABLE json_nested_test ( count string, usage string, pkg map<string

hive報錯 Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:For direct MetaStore DB connections,

pre mysq rom ges character base ddltask for latin 學習hive 使用mysql作為元數據 hive創建數據庫和切換數據庫都是可以的 但是創建表就是出問題 百度之後發現 是編碼問題 特別記錄一下~~~ 1.報錯前如圖:

org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does not exist: file:/input

utf test exceptio 執行 cep exc 文件目錄 XML 配置 原我是這樣寫的 //輸入數據所在的文件目錄 FileInputFormat.addInputPath(job, new Path("/input/")); //mapreduce執行後

配置MapReduce插件時,彈窗報錯org/apache/hadoop/eclipse/preferences/MapReducePreferencePage : Unsupported major.minor version 51.0(Hadoop2.7.3集群部署)

ava 不一致 nbsp 1.0 log class dll blog 無效 原因: hadoop-eclipse-plugin-2.7.3.jar 編譯的jdk版本和eclipse啟動使用的jdk版本不一致導致。 解決方案一: 修改myeclipse.ini文件

shematool -initschema -dbtype mysql error org.apache.hadoop.hive.metastore.hivemetaexception:Failed to get schema version

hang my.cnf blog address com rest chang init edit 命令:schematool -initSchema -dbType mysql Fix the issue: edit /etc/mysql/my.cnf change b

java.lang.NullPointerException at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:482)

想要 fileutil acc 測試 cep .net parent int pre 1:問題出現的原因,部署好的hadoop-2.6.4進行window10操作hadoop api出現的錯誤,具體錯誤是我向hdfs上傳文件,還好點,之前解決過,這裏不敘述,這裏說一下從hd