hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Mahadev konar (JIRA)" <j...@apache.org>
Subject [jira] Created: (HADOOP-2955) ant test fail for TestCrcCorruption with OutofMemory.
Date Thu, 06 Mar 2008 19:30:57 GMT
ant test fail for TestCrcCorruption with OutofMemory.

                 Key: HADOOP-2955
                 URL: https://issues.apache.org/jira/browse/HADOOP-2955
             Project: Hadoop Core
          Issue Type: Bug
            Reporter: Mahadev konar
            Assignee: Raghu Angadi
            Priority: Blocker

TestCrcCorruption sometimes corrupts the metadata for crc and leads to corruption in the length
of of bytes of checksum (second field in metadata). This does not happen always but somtimes
since corruption is random in the test.

I put in a debug statement in the allocation to see how many bytes were being allocated and
ran it for few times. This is one of the allocation in 

 int maxChunksPerPacket = Math.max(1,
                      (BUFFER_SIZE + bytesPerChecksum - 1)/bytesPerChecksum);
        int sizeofPacket = PKT_HEADER_LEN + 
        (bytesPerChecksum + checksumSize) * maxChunksPerPacket;
        LOG.info("Comment: bytes to allocate " + sizeofPacket);
        ByteBuffer pktBuf = ByteBuffer.allocate(sizeofPacket);

The output in one of the allocations is 

 dfs.DataNode (DataNode.java:sendBlock(1766)) - Comment: bytes to allocate 1232596786

So we should check for number of bytes being allocated in sendBlock (should be less than the
block size? -- seems like a good default).

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message