spark on yarn 报 org.apache.hadoop.util.Shell$ExitCodeException: 问题

在将基于Spark 1.6、JDK 1.7和Hadoop 2.7的Scala程序移植到使用JDK 1.6、Hadoop 2.2的环境中进行Spark on YARN测试时,遇到`org.apache.hadoop.util.Shell$ExitCodeException`错误。尝试更新jar、driver-class-path等方式未解决问题。通过分析日志发现,问题可能与JDK版本有关。实验表明,使用JDK 1.7的NodeManager可以成功运行,而JDK 1.6的则失败。最终,通过降级到Spark 1.4.0成功解决了问题,证实问题确实由版本不兼容引起。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

今天把新数据中心基于 spark 1.6 jdk 1.7 hadoop2.7的scala程序 拿到老数据中心jdk1.6/jdk.17 hadoop 2.2 环境 进行 spark on yarn测试
使用spark-shell 或者 spark-submit 提交时报如下错误

org.apache.hadoop.util.Shell$ExitCodeException: 
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
        at org.apache.hadoop.util.Shell.run(Shell.java:379)
        at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
        at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
        at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283)
        at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79)
        at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
        at java.util.concurrent.FutureTask.run(FutureTask.java:138)
        at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
        at java.lang.Thread.run(Thread.java:662)


.Failing this attempt.. Failing the application.
         ApplicationMaster host: N/A
         ApplicationMaster RPC port: 0
         queue: default
         start time: 1481532476948
         final status: FAILED
         tracking URL: xxx:8088/cluster/app/application_1480324693568_0053
Exception in thread "main" org.apache.spark.SparkException: Yarn application has already ended! It might have been killed or unable to launch application master.
        at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:124)
        at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:64)
        at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:144)
        at org.apache.spark.SparkContext.<init>(SparkContext.scala:530)
        at Regex$.main(Regex.scala:14)
        at Regex.main(Regex.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:483)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)

查看container日志 报如下错误

xxx(&container_1480324693568_0052_01_000001έstderr661Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/spark/deploy/yarn/ExecutorLauncher
Caused by: java.lang.ClassNotFoundException: org.apache.spark.deploy.yarn.ExecutorLauncher
    at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
    at java.security.AccessController.doPrivileged(Native Method)
    at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:247)

一开始以为是 spark基于hadoop的包spark_home/lib/spark-assembly-1.6.1-hadoop2.3.0.jar 没有加载 还试着用–jar 或者 –diver-class-path 加载 但是后来发现了启动客户端日志里面 有这句话

16/12/12 16:47:54 INFO yarn.Client: Uploading resource file:/home/xxx/spark-1.6.1/lib/spark-assembly-1.6.1-hadoop2.3.0.jar -> hdfs://ns1/user/xxx/.sparkStaging/application_1480324693568_0053/spark-assembly-1.6.1-hadoop2.3.0.jar

说明该工具包已经上送的yarn环境 还怀疑这个类 不在这个包里 但是通过eclipse 反查类 这个类是在里面的
比较巧合的是 程序 在 有两个节点运行成功了 后来发现 这两个节点的nodemanager是jdk1.7运行的 而其他的 失败的节点是 jdk1.6运行的 所以怀疑是jdk版本早造成的 所以 换这个环境试试spark-1.4.0-bin-hadoop2.3
最后成功
看来还是版本的问题

INFO Client:54 - client token: N/A diagnostics: Application application_1686359558924_0001 failed 2 times due to AM Container for appattempt_1686359558924_0001_000002 exited with exitCode: 13 For more detailed output, check application tracking page:http://spark01:8088/cluster/app/application_1686359558924_0001Then, click on links to logs of each attempt. Diagnostics: Exception from container-launch. Container id: container_e01_1686359558924_0001_02_000001 Exit code: 13 Stack trace: ExitCodeException exitCode=13: at org.apache.hadoop.util.Shell.runCommand(Shell.java:585) at org.apache.hadoop.util.Shell.run(Shell.java:482) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:776) at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:212) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Container exited with a non-zero exit code 13 Failing this attempt. Failing the application. ApplicationMaster host: N/A ApplicationMaster RPC port: -1 queue: default start time: 1686359786278 final status: FAILED tracking URL: http://spark01:8088/cluster/app/application_1686359558924_0001 user: root Exception in thread "main" org.apache.spark.SparkException: Application application_1686359558924_0001 finished with failed status at org.apache.spark.deploy.yarn.Client.run(Client.scala:1165) at org.apache.spark.deploy.yarn.YarnClusterApplication.start(Client.scala:1520) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2023-06-10 09:20:01 INFO ShutdownHookManager:54 - Shutdown hook called 2023-06-10 09:20:01 INFO ShutdownHookManager:54 - Deleting directory /tmp/spark-7a611aba-e3f1-4ee9-9829-f1bc45c9348d 2023-06-10 09:20:01 INFO ShutdownHookManager:54 - Deleting directory /tmp/spark-b69c8ca8-924f-430e-9a7c-625efd9e9f47
06-11
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值