spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "yao zhang (JIRA)" <>
Subject [jira] [Commented] (SPARK-20579) large spark job hang on with many active stages/jobs
Date Wed, 03 May 2017 13:51:04 GMT


yao zhang commented on SPARK-20579:

I recently use scala spark to do complex data processing on large data set, but always run
into hanged job with many active jobs/stages in web UI.
Here is details:  
   1. it only happens in spark-submit, when I test code in spark-shell, I never has this issue.
   2. I always use dynamic allocation. 
   3. My cluster is large enough for my task (I can use 2000 executors with mem 8G, driver
with mem 100G).
   4. The basic issue pattern is (from UI): some stages incomplete, then always keep active
=> corresponding jobs incomplete and always active => active tasks in executors accumulate
=> RDD blocks in executors accumulate => executors get locked => application hanged
and cannot move 

> large spark job hang on with many active stages/jobs
> ----------------------------------------------------
>                 Key: SPARK-20579
>                 URL:
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Submit
>    Affects Versions: 2.1.0
>         Environment: spark 2.10 in hadoop (2.6.0) cluster
>            Reporter: yao zhang
>         Attachments: executor-screen.png, job-screen.png, stage-screen.png, storage-screen.png,

This message was sent by Atlassian JIRA

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message