hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Raghu Angadi (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-6460) Namenode runs of out of memory due to memory leak in ipc Server
Date Mon, 21 Dec 2009 22:35:18 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-6460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12793394#action_12793394
] 

Raghu Angadi commented on HADOOP-6460:
--------------------------------------

Is the space taken by a few very large buffers (on the order of number of handlers) or large
number of smaller buffers?

'reset()' does not change the underlying array size, it keeps growing if you have larger and
larger responses to RPCs. 


> Namenode runs of out of memory due to memory leak in ipc Server
> ---------------------------------------------------------------
>
>                 Key: HADOOP-6460
>                 URL: https://issues.apache.org/jira/browse/HADOOP-6460
>             Project: Hadoop Common
>          Issue Type: Bug
>    Affects Versions: 0.20.1, 0.21.0, 0.22.0
>            Reporter: Suresh Srinivas
>            Assignee: Suresh Srinivas
>            Priority: Blocker
>             Fix For: 0.20.2, 0.21.0, 0.22.0
>
>
> Namenode heap usage grows disproportional to the number objects supports (files, directories
and blocks). Based on heap dump analysis, this is due to large growth in ByteArrayOutputStream
allocated in o.a.h.ipc.Server.Handler.run().

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message