hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Enis Soztutar (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-2569) Unit test times out on Solaris nightly build: mapred.TestMultiFileInputFormat
Date Thu, 10 Jan 2008 07:50:34 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-2569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12557563#action_12557563
] 

Enis Soztutar commented on HADOOP-2569:
---------------------------------------

Well it seems to have stuck in lines below because the log after calling initFiles() is not
printed. 
{code}
   LOG.info("Creating " + numFiles + " file(s) in " + multiFileDir);
    for(int i=0; i<numFiles ;i++) {
      Path path = new Path(multiFileDir, "file_" + i);
       FSDataOutputStream out = fs.create(path);
       if (numBytes == -1) {
         numBytes = rand.nextInt(MAX_BYTES);
       }
       for(int j=0; j< numBytes; j++) {
         out.write(rand.nextInt());
       }
       out.close();
       if(LOG.isDebugEnabled()) {
         LOG.debug("Created file " + path + " with length " + numBytes);
       }
       lengths.put(path.getName(), new Long(numBytes));
    }
{code}

I assume there is some space related problem in the build server. The test log reports many
other problems such as : 
{noformat}
...
[junit] 2008-01-09 11:35:47,543 WARN  dfs.DataNode (DataNode.java:copyBlock(1144)) - Got exception
while serving blk_-4990242691406600933 to 127.0.0.1:35838: java.io.IOException: Block blk_-4990242691406600933
is not valid.
    [junit] 	at org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:549)
    [junit] 	at org.apache.hadoop.dfs.FSDataset.getMetaFile(FSDataset.java:466)
    [junit] 	at org.apache.hadoop.dfs.FSDataset.getMetaDataInputStream(FSDataset.java:480)
    [junit] 	at org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1298)
    [junit] 	at org.apache.hadoop.dfs.DataNode$DataXceiver.copyBlock(DataNode.java:1114)
    [junit] 	at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:877)
    [junit] 	at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 11:35:56,574 ERROR dfs.DataNode (DataNode.java:run(1738)) - Exception:
java.lang.reflect.UndeclaredThrowableException
    [junit] 	at org.apache.hadoop.dfs.$Proxy1.blockReport(Unknown Source)
    [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:616)
    [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1736)
    [junit] 	at java.lang.Thread.run(Thread.java:595)
    [junit] Caused by: java.lang.InterruptedException
    [junit] 	at java.lang.Object.wait(Native Method)
    [junit] 	at org.apache.hadoop.ipc.Client.call(Client.java:504)
    [junit] 	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
    [junit] 	... 4 more
...
[junit] 2008-01-09 11:37:31,788 WARN  dfs.DataNode (DataNode.java:sendChunk(1401)) -  Could
not read checksum for data at offset 1536 for block blk_-5451683627314743712 got : java.io.EOFException
    [junit] 	at java.io.DataInputStream.readFully(DataInputStream.java:178)
    [junit] 	at org.apache.hadoop.dfs.DataNode$BlockSender.sendChunk(DataNode.java:1399)
    [junit] 	at org.apache.hadoop.dfs.DataNode$BlockSender.sendBlock(DataNode.java:1449)
    [junit] 	at org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:920)
    [junit] 	at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:865)
    [junit] 	at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 12:36:51,570 WARN  fs.AllocatorPerContext (LocalDirAllocator.java:createPath(256))
- org.apache.hadoop.util.DiskChecker$DiskErrorException: directory is not writable: build/test/temp/tmp4
    [junit] 	at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:85)
    [junit] 	at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createPath(LocalDirAllocator.java:253)
    [junit] 	at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:301)
    [junit] 	at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWrite(LocalDirAllocator.java:326)
    [junit] 	at org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator.java:155)
    [junit] 	at org.apache.hadoop.fs.TestLocalDirAllocator.createTempFile(TestLocalDirAllocator.java:77)
...
{noformat}

can you check for the server whether there is a space problem or the directories are writable
(I don't have a hudson account). 


> Unit test times out on Solaris nightly build: mapred.TestMultiFileInputFormat
> -----------------------------------------------------------------------------
>
>                 Key: HADOOP-2569
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2569
>             Project: Hadoop
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.16.0
>         Environment: solaris
>            Reporter: Mukund Madhugiri
>            Assignee: Enis Soztutar
>            Priority: Blocker
>             Fix For: 0.16.0
>
>
> Unit test failed in the nightly build: org.apache.hadoop.mapred.TestMultiFileInputFormat.unknown
> junit.framework.AssertionFailedError: Timeout occurred
> Logs are at:
> http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Nightly/361/

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message