hadoop-yarn-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (YARN-2230) Fix description of yarn.scheduler.maximum-allocation-vcores in yarn-default.xml (or code)
Date Wed, 17 Dec 2014 19:15:14 GMT

    [ https://issues.apache.org/jira/browse/YARN-2230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14250340#comment-14250340

Hadoop QA commented on YARN-2230:

{color:red}-1 overall{color}.  Here are the results of testing the latest attachment 
  against trunk revision 4281c96.

    {color:green}+1 @author{color}.  The patch does not contain any @author tags.

    {color:red}-1 tests included{color}.  The patch doesn't appear to include any new or modified
                        Please justify why no new tests are needed for this patch.
                        Also please list what manual steps were performed to verify this patch.

    {color:green}+1 javac{color}.  The applied patch does not increase the total number of
javac compiler warnings.

    {color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

    {color:green}+1 eclipse:eclipse{color}.  The patch built with eclipse:eclipse.

    {color:red}-1 findbugs{color}.  The patch appears to introduce 25 new Findbugs (version
2.0.3) warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase the total number
of release audit warnings.

    {color:green}+1 core tests{color}.  The patch passed unit tests in hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

Test results: https://builds.apache.org/job/PreCommit-YARN-Build/6131//testReport/
Findbugs warnings: https://builds.apache.org/job/PreCommit-YARN-Build/6131//artifact/patchprocess/newPatchFindbugsWarningshadoop-yarn-common.html
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/6131//console

This message is automatically generated.

> Fix description of yarn.scheduler.maximum-allocation-vcores in yarn-default.xml (or code)
> -----------------------------------------------------------------------------------------
>                 Key: YARN-2230
>                 URL: https://issues.apache.org/jira/browse/YARN-2230
>             Project: Hadoop YARN
>          Issue Type: Bug
>          Components: client, scheduler
>    Affects Versions: 2.4.0
>            Reporter: Adam Kawa
>            Priority: Minor
>         Attachments: YARN-2230.001.patch
> When a user requests more vcores than the allocation limit (e.g. mapreduce.map.cpu.vcores
 is larger than yarn.scheduler.maximum-allocation-vcores), then InvalidResourceRequestException
is thrown - https://svn.apache.org/repos/asf/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerUtils.java
> {code}
>     if (resReq.getCapability().getVirtualCores() < 0 ||
>         resReq.getCapability().getVirtualCores() >
>         maximumResource.getVirtualCores()) {
>       throw new InvalidResourceRequestException("Invalid resource request"
>           + ", requested virtual cores < 0"
>           + ", or requested virtual cores > max configured"
>           + ", requestedVirtualCores="
>           + resReq.getCapability().getVirtualCores()
>           + ", maxVirtualCores=" + maximumResource.getVirtualCores());
>     }
> {code}
> According to documentation - yarn-default.xml http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-common/yarn-default.xml,
the request should be capped to the allocation limit.
> {code}
>   <property>
>     <description>The maximum allocation for every container request at the RM,
>     in terms of virtual CPU cores. Requests higher than this won't take effect,
>     and will get capped to this value.</description>
>     <name>yarn.scheduler.maximum-allocation-vcores</name>
>     <value>32</value>
>   </property>
> {code}
> This means that:
> * Either documentation or code should be corrected (unless this exception is handled
elsewhere accordingly, but it looks that it is not).
> This behavior is confusing, because when such a job (with mapreduce.map.cpu.vcores is
larger than yarn.scheduler.maximum-allocation-vcores) is submitted, it does not make any progress.
The warnings/exceptions are thrown at the scheduler (RM) side e.g.
> {code}
> 2014-06-29 00:34:51,469 WARN org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService:
Invalid resource ask by application appattempt_1403993411503_0002_000001
> org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: Invalid resource request,
requested virtual cores < 0, or requested virtual cores > max configured, requestedVirtualCores=32,
> 	at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:237)
> 	at org.apache.hadoop.yarn.server.resourcemanager.RMServerUtils.validateResourceRequests(RMServerUtils.java:80)
> 	at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:420)
>         .....
> 	at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
> 	at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1026)
> 	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1986)
> 	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1982)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:416)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
> 	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1980)
> {code}
> * IMHO, such an exception should be forwarded to client. Otherwise, it is non obvious
to discover why a job does not make any progress.
> The same looks to be related to memory.

This message was sent by Atlassian JIRA

View raw message