OOZIE:JA009:RPC响应超出最大数据长度

时间:2018-12-05 13:06:48

标签: hadoop mapreduce sqoop oozie

OOZIE字数示例给出了JA009:RPC响应超出了最大数据长度。我们已经将ipc.maximum.data.length加倍,并重新启动了NameNode。

2018-12-05 17:55:45,914  WARN MapReduceActionExecutor:523 - SERVER[******] USER[******] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000004-181205174411487-oozie-******-W] ACTION[0000004-181205174411487-oozie-******-W@mr-node] No credential properties found for action : 0000004-181205174411487-oozie-******-W@mr-node, cred : null
2018-12-05 18:10:46,019  WARN ActionStartXCommand:523 - SERVER[******] USER[******] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000004-181205174411487-oozie-******-W] ACTION[0000004-181205174411487-oozie-******-W@mr-node] Error starting action [mr-node]. ErrorType [TRANSIENT], ErrorCode [JA009], Message [JA009: RPC response exceeds maximum data length]
org.apache.oozie.action.ActionExecutorException: JA009: RPC response exceeds maximum data length
    at org.apache.oozie.action.ActionExecutor.convertExceptionHelper(ActionExecutor.java:463)
    at org.apache.oozie.action.ActionExecutor.convertException(ActionExecutor.java:437)
    at org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:1070)
    at org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:1512)
    at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:243)
    at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:68)
    at org.apache.oozie.command.XCommand.call(XCommand.java:290)
    at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:334)
    at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:263)
    at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:181)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.ipc.RpcException: RPC response exceeds maximum data length
    at org.apache.hadoop.ipc.Client$IpcStreams.readResponse(Client.java:1808)
    at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1163)
    at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1059)

任何帮助将不胜感激。谢谢

1 个答案:

答案 0 :(得分:0)

您是否尝试修改hdfs-site.xml中的配置?

<property>
     <name>ipc.maximum.data.length</name>
     <value>134217728</value>
</property>

如果已经足够高,请确保fs.default.name的core-site.xml配置使用的是IP,而不仅仅是本地主机。

<configuration>
       ....
        <property>
                <name>fs.default.name</name>
                <value>hdfs://your ip:9000</value>
        </property>
</configuration>