hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-8361) avoid out-of-memory problems when deserializing strings
Date Tue, 08 May 2012 00:11:51 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-8361?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13270102#comment-13270102

Hadoop QA commented on HADOOP-8361:

+1 overall.  Here are the results of testing the latest attachment 
  against trunk revision .

    +1 @author.  The patch does not contain any @author tags.

    +1 tests included.  The patch appears to include 1 new or modified test files.

    +1 javadoc.  The javadoc tool did not generate any warning messages.

    +1 javac.  The applied patch does not increase the total number of javac compiler warnings.

    +1 eclipse:eclipse.  The patch built with eclipse:eclipse.

    +1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) warnings.

    +1 release audit.  The applied patch does not increase the total number of release audit

    +1 core tests.  The patch passed unit tests in hadoop-common-project/hadoop-common.

    +1 contrib tests.  The patch passed contrib unit tests.

Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/956//testReport/
Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/956//console

This message is automatically generated.
> avoid out-of-memory problems when deserializing strings
> -------------------------------------------------------
>                 Key: HADOOP-8361
>                 URL: https://issues.apache.org/jira/browse/HADOOP-8361
>             Project: Hadoop Common
>          Issue Type: Bug
>            Reporter: Colin Patrick McCabe
>            Assignee: Colin Patrick McCabe
>            Priority: Minor
>         Attachments: HADOOP-8361.001.patch, HADOOP-8361.002.patch, HADOOP-8361.003.patch,
HADOOP-8361.004.patch, HADOOP-8361.005.patch
> In HDFS, we want to be able to read the edit log without crashing on an OOM condition.
 Unfortunately, we currently cannot do this, because there are no limits on the length of
certain data types we pull from the edit log.  We often read strings without setting any upper
limit on the length we're prepared to accept.
> It's not that we don't have limits on strings-- for example, HDFS limits the maximum
path length to 8000 UCS-2 characters.  Linux limits the maximum user name length to either
64 or 128 bytes, depending on what version you are running.  It's just that we're not exposing
these limits to the deserialization functions that need to be aware of them.

This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira


View raw message