spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Apache Spark (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (SPARK-23660) Yarn throws exception in cluster mode when the application is small
Date Tue, 13 Mar 2018 04:47:00 GMT

    [ https://issues.apache.org/jira/browse/SPARK-23660?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16396541#comment-16396541
] 

Apache Spark commented on SPARK-23660:
--------------------------------------

User 'gaborgsomogyi' has created a pull request for this issue:
https://github.com/apache/spark/pull/20807

> Yarn throws exception in cluster mode when the application is small
> -------------------------------------------------------------------
>
>                 Key: SPARK-23660
>                 URL: https://issues.apache.org/jira/browse/SPARK-23660
>             Project: Spark
>          Issue Type: Bug
>          Components: YARN
>    Affects Versions: 2.3.0
>            Reporter: Gabor Somogyi
>            Priority: Minor
>
> Yarn throws the following exception in cluster mode when the application is really small:
> {code:java}
> 18/03/07 23:34:22 WARN netty.NettyRpcEnv: Ignored failure: java.util.concurrent.RejectedExecutionException:
Task java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@7c974942 rejected
from java.util.concurrent.ScheduledThreadPoolExecutor@1eea9d2d[Terminated, pool size = 0,
active threads = 0, queued tasks = 0, completed tasks = 0]
> 18/03/07 23:34:22 ERROR yarn.ApplicationMaster: Uncaught exception: 
> org.apache.spark.SparkException: Exception thrown in awaitResult: 
> 	at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205)
> 	at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
> 	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)
> 	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)
> 	at org.apache.spark.deploy.yarn.YarnAllocator.<init>(YarnAllocator.scala:102)
> 	at org.apache.spark.deploy.yarn.YarnRMClient.register(YarnRMClient.scala:77)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.registerAM(ApplicationMaster.scala:450)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:493)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:345)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply$mcV$sp(ApplicationMaster.scala:260)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$5.run(ApplicationMaster.scala:810)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:809)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:259)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:834)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
> Caused by: org.apache.spark.rpc.RpcEnvStoppedException: RpcEnv already stopped.
> 	at org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:158)
> 	at org.apache.spark.rpc.netty.Dispatcher.postLocalMessage(Dispatcher.scala:135)
> 	at org.apache.spark.rpc.netty.NettyRpcEnv.ask(NettyRpcEnv.scala:229)
> 	at org.apache.spark.rpc.netty.NettyRpcEndpointRef.ask(NettyRpcEnv.scala:523)
> 	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:91)
> 	... 17 more
> 18/03/07 23:34:22 INFO yarn.ApplicationMaster: Final app status: FAILED, exitCode: 13,
(reason: Uncaught exception: org.apache.spark.SparkException: Exception thrown in awaitResult:
)
> {code}
> Example application:
> {code:java}
> object ExampleApp {
>   def main(args: Array[String]): Unit = {
>     val conf = new SparkConf().setAppName("ExampleApp")
>     val sc = new SparkContext(conf)
>     try {
>       // Do nothing
>     } finally {
>       sc.stop()
>     }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org


Mime
View raw message