hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Raghu Angadi (JIRA)" <j...@apache.org>
Subject [jira] Updated: (HADOOP-1035) StackOverflowError in FSDataSet
Date Wed, 28 Feb 2007 01:42:06 GMT

     [ https://issues.apache.org/jira/browse/HADOOP-1035?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Raghu Angadi updated HADOOP-1035:
---------------------------------

    Attachment: HADOOP-1035-1.patch


1035-1.patch is attached. Here addBlock's recursion is limited to depth of the directory tree.
Also we now create children under all the sub directories and this will keep tree depth in
check.

Each directory remembers last child that had room for a new block. This way most of the calls
to addBlock would only traverses one line in the tree.

clearPath() also had equally had recursion. This is fixed too.

Phillippe, please take look at the patch and use it if you feel confident about it.


> StackOverflowError in FSDataSet
> -------------------------------
>
>                 Key: HADOOP-1035
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1035
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.10.1, 0.11.0, 0.11.1, 0.11.2
>            Reporter: Philippe Gassmann
>         Assigned To: Raghu Angadi
>            Priority: Blocker
>         Attachments: HADOOP-1035-1.patch, patch-StackOverflowError-HADOOP-1035
>
>
> [hadoop.org.apache.hadoop.dfs.DataNode] DataXCeiver
> java.lang.StackOverflowError
> at java.nio.ByteBuffer.wrap([BII)Ljava.nio.ByteBuffer;(Unknown Source)
> at java.nio.ByteBuffer.wrap([B)Ljava.nio.ByteBuffer;(Unknown Source)
> at java.lang.StringCoding$CharsetSE.encode([CII)[B(Unknown Source)
> at java.lang.StringCoding.encode(Ljava.lang.String;[CII)[B(Unknown Source)
> at java.lang.String.getBytes(Ljava.lang.String;)[B(Unknown Source)
> at java.io.UnixFileSystem.rename0(Ljava.io.File;Ljava.io.File;)Z(Native Method)
> at java.io.UnixFileSystem.rename(UnixFileSystem.java:265)
> at java.io.File.renameTo(File.java:1192)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:89)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:105)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> at org.apache.hadoop.dfs.FSDataset$FSDir.addBlock(FSDataset.java:95)
> I do not have the end of the stacktrace, but it is sure it happens in DataNode.DataXceiver.writeBlock().
> This error occurs after applying the patch provided in HADOOP-1034 that permits to see
such exceptions in log files.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message