hadoop-hdfs-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Hudson Server <hud...@hudson.apache.org>
Subject Hadoop-Hdfs-trunk - Build # 508 - Still Failing
Date Thu, 02 Dec 2010 16:31:29 GMT
See https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/508/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 855880 lines...]
    [junit] 2010-12-02 16:33:27,065 INFO  net.NetworkTopology (NetworkTopology.java:add(331))
- Adding a new node: /default-rack/127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,069 INFO  datanode.DataNode (DataNode.java:register(697))
- New storage id DS-1896102694-127.0.1.1-34336-1291307607063 is assigned to data-node 127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,070 INFO  datanode.DataNode (DataNode.java:run(1421)) - DatanodeRegistration(127.0.0.1:34336,
storageID=DS-1896102694-127.0.1.1-34336-1291307607063, infoPort=51649, ipcPort=59557)In DataNode.run,
data = FSDataset{dirpath='/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data5/current/finalized,/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data6/current/finalized'}
    [junit] 2010-12-02 16:33:27,070 INFO  ipc.Server (Server.java:run(608)) - IPC Server Responder:
starting
    [junit] 2010-12-02 16:33:27,070 INFO  datanode.DataNode (DataNode.java:offerService(887))
- using BLOCKREPORT_INTERVAL of 21600000msec Initial delay: 0msec
    [junit] 2010-12-02 16:33:27,070 INFO  ipc.Server (Server.java:run(1369)) - IPC Server
handler 0 on 59557: starting
    [junit] 2010-12-02 16:33:27,070 INFO  ipc.Server (Server.java:run(443)) - IPC Server listener
on 59557: starting
    [junit] 2010-12-02 16:33:27,076 INFO  datanode.DataNode (DataNode.java:blockReport(1126))
- BlockReport of 0 blocks got processed in 2 msecs
    [junit] 2010-12-02 16:33:27,076 INFO  datanode.DataNode (DataNode.java:offerService(929))
- Starting Periodic block scanner.
    [junit] 2010-12-02 16:33:27,088 INFO  FSNamesystem.audit (FSNamesystem.java:logAuditEvent(148))
- ugi=hudson	ip=/127.0.0.1	cmd=create	src=/pipeline_Fi_43/foo	dst=null	perm=hudson:supergroup:rw-r--r--
    [junit] 2010-12-02 16:33:27,089 INFO  hdfs.DFSClientAspects (DFSClientAspects.aj:ajc$before$org_apache_hadoop_hdfs_DFSClientAspects$5$5ba7280d(86))
- FI: before pipelineClose:
    [junit] 2010-12-02 16:33:27,091 INFO  hdfs.StateChange (FSNamesystem.java:allocateBlock(1753))
- BLOCK* NameSystem.allocateBlock: /pipeline_Fi_43/foo. blk_-5987900848781850642_1001{blockUCState=UNDER_CONSTRUCTION,
primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[127.0.0.1:53619|RBW], ReplicaUnderConstruction[127.0.0.1:34336|RBW],
ReplicaUnderConstruction[127.0.0.1:39706|RBW]]}
    [junit] 2010-12-02 16:33:27,092 INFO  protocol.ClientProtocolAspects (ClientProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_protocol_ClientProtocolAspects$1$7076326d(35))
- FI: addBlock Pipeline[127.0.0.1:53619, 127.0.0.1:34336, 127.0.0.1:39706]
    [junit] 2010-12-02 16:33:27,093 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,093 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-12-02 16:33:27,093 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(257))
- Receiving block blk_-5987900848781850642_1001 src: /127.0.0.1:35600 dest: /127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,095 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,095 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-12-02 16:33:27,095 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(257))
- Receiving block blk_-5987900848781850642_1001 src: /127.0.0.1:54758 dest: /127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,097 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,097 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-12-02 16:33:27,097 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(257))
- Receiving block blk_-5987900848781850642_1001 src: /127.0.0.1:50309 dest: /127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,098 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
- FI: statusRead SUCCESS, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,098 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,098 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,098 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
- FI: statusRead SUCCESS, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,099 INFO  hdfs.DFSClientAspects (DFSClientAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_DFSClientAspects$2$9396d2df(48))
- FI: after pipelineInitNonAppend: hasError=false errorIndex=-1
    [junit] 2010-12-02 16:33:27,099 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,100 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,100 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
- FI: callWritePacketToDisk
    [junit] 2010-12-02 16:33:27,100 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,100 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,100 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
- FI: callWritePacketToDisk
    [junit] 2010-12-02 16:33:27,101 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,101 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
- FI: callWritePacketToDisk
    [junit] 2010-12-02 16:33:27,101 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,101 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$after$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$8$9594fb70(205))
- FI: fiPipelineAck, datanode=DatanodeRegistration(127.0.0.1:34336, storageID=DS-1896102694-127.0.1.1-34336-1291307607063,
infoPort=51649, ipcPort=59557)
    [junit] 2010-12-02 16:33:27,101 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,102 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$after$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$8$9594fb70(205))
- FI: fiPipelineAck, datanode=DatanodeRegistration(127.0.0.1:53619, storageID=DS-2014057751-127.0.1.1-53619-1291307606796,
infoPort=47703, ipcPort=49272)
    [junit] 2010-12-02 16:33:27,102 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:53619
    [junit] 2010-12-02 16:33:27,102 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$7$b9c2bffe(184))
- FI: pipelineClose, datanode=127.0.0.1:53619, offsetInBlock=1, seqno=1, lastPacketInBlock=true,
len=0, endOfHeader=25
    [junit] 2010-12-02 16:33:27,103 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:34336
    [junit] 2010-12-02 16:33:27,103 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$7$b9c2bffe(184))
- FI: pipelineClose, datanode=127.0.0.1:34336, offsetInBlock=1, seqno=1, lastPacketInBlock=true,
len=0, endOfHeader=25
    [junit] 2010-12-02 16:33:27,103 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,103 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$7$b9c2bffe(184))
- FI: pipelineClose, datanode=127.0.0.1:39706, offsetInBlock=1, seqno=1, lastPacketInBlock=true,
len=0, endOfHeader=25
    [junit] 2010-12-02 16:33:27,104 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_43, index=2, duration=[3000, 3001), datanode=127.0.0.1:39706
    [junit] 2010-12-02 16:33:27,104 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:50309 [Receiving block blk_-5987900848781850642_1001 client=DFSClient_804440136]
sleeps for 3000ms
Build timed out. Aborting
[FINDBUGS] Skipping publisher since build result is FAILURE
Publishing Javadoc
Archiving artifacts
Recording test results
Recording fingerprints
Error updating JIRA issues. Saving issues for next build.
com.atlassian.jira.rpc.exception.RemotePermissionException: This issue does not exist or you
don't have permission to view it.
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
15 tests failed.
FAILED:  TEST-org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol.xml.<init>

Error Message:


Stack Trace:
Test report file /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/TEST-org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol.xml
was length 0

REGRESSION:  org.apache.hadoop.hdfs.server.datanode.TestBlockReport.blockReport_08

Error Message:
Was waiting too long for a replica to become TEMPORARY

Stack Trace:
junit.framework.AssertionFailedError: Was waiting too long for a replica to become TEMPORARY
	at org.apache.hadoop.hdfs.server.datanode.TestBlockReport.waitForTempReplica(TestBlockReport.java:514)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockReport.__CLR3_0_2j2e00jqbq(TestBlockReport.java:408)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockReport.blockReport_08(TestBlockReport.java:390)


FAILED:  org.apache.hadoop.hdfs.TestFileAppend4.testRecoverFinalizedBlock

Error Message:
test timed out after 60000 milliseconds

Stack Trace:
java.lang.Exception: test timed out after 60000 milliseconds
	at java.io.FileInputStream.readBytes(Native Method)
	at java.io.FileInputStream.read(FileInputStream.java:199)
	at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
	at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
	at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
	at java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
	at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
	at sun.security.provider.SeedGenerator$URLSeedGenerator.getSeedByte(SeedGenerator.java:453)
	at sun.security.provider.SeedGenerator.getSeedBytes(SeedGenerator.java:123)
	at sun.security.provider.SeedGenerator.generateSeed(SeedGenerator.java:118)
	at sun.security.provider.SecureRandom.engineGenerateSeed(SecureRandom.java:114)
	at sun.security.provider.SecureRandom.engineNextBytes(SecureRandom.java:171)
	at java.security.SecureRandom.nextBytes(SecureRandom.java:433)
	at java.security.SecureRandom.next(SecureRandom.java:455)
	at java.util.Random.nextLong(Random.java:284)
	at org.mortbay.jetty.servlet.HashSessionIdManager.doStart(HashSessionIdManager.java:139)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.mortbay.jetty.servlet.AbstractSessionManager.doStart(AbstractSessionManager.java:168)
	at org.mortbay.jetty.servlet.HashSessionManager.doStart(HashSessionManager.java:67)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.mortbay.jetty.servlet.SessionHandler.doStart(SessionHandler.java:115)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
	at org.mortbay.jetty.handler.ContextHandler.startContext(ContextHandler.java:537)
	at org.mortbay.jetty.servlet.Context.startContext(Context.java:136)
	at org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1234)
	at org.mortbay.jetty.handler.ContextHandler.doStart(ContextHandler.java:517)
	at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:460)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
	at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
	at org.mortbay.jetty.Server.doStart(Server.java:222)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	at org.apache.hadoop.http.HttpServer.start(HttpServer.java:618)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:516)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:461)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1115)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:461)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.activate(NameNode.java:405)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:389)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:578)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:571)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1534)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:445)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:176)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:71)
	at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:168)
	at org.apache.hadoop.hdfs.TestFileAppend4.__CLR3_0_21z1ppcxud(TestFileAppend4.java:151)
	at org.apache.hadoop.hdfs.TestFileAppend4.testRecoverFinalizedBlock(TestFileAppend4.java:150)


FAILED:  org.apache.hadoop.hdfs.TestFileAppend4.testCompleteOtherLeaseHoldersFile

Error Message:
Cannot lock storage /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/name1.
The directory is already locked.

Stack Trace:
java.io.IOException: Cannot lock storage /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/name1.
The directory is already locked.
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:615)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1332)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1350)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1408)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:202)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:435)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:176)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:71)
	at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:168)
	at org.apache.hadoop.hdfs.TestFileAppend4.__CLR3_0_269ddf9xvh(TestFileAppend4.java:222)
	at org.apache.hadoop.hdfs.TestFileAppend4.testCompleteOtherLeaseHoldersFile(TestFileAppend4.java:221)


FAILED:  org.apache.hadoop.hdfs.TestFileConcurrentReader.testUnfinishedBlockCRCErrorTransferToVerySmallWrite

Error Message:
java.io.FileNotFoundException: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/classes/hdfs-default.xml
(Too many open files)

Stack Trace:
java.lang.RuntimeException: java.io.FileNotFoundException: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/classes/hdfs-default.xml
(Too many open files)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1523)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:1388)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:1334)
	at org.apache.hadoop.conf.Configuration.set(Configuration.java:577)
	at org.apache.hadoop.conf.Configuration.setBoolean(Configuration.java:781)
	at org.apache.hadoop.hdfs.TestFileConcurrentReader.runTestUnfinishedBlockCRCError(TestFileConcurrentReader.java:313)
	at org.apache.hadoop.hdfs.TestFileConcurrentReader.runTestUnfinishedBlockCRCError(TestFileConcurrentReader.java:302)
	at org.apache.hadoop.hdfs.TestFileConcurrentReader.__CLR3_0_2u5mf5tqqo(TestFileConcurrentReader.java:275)
	at org.apache.hadoop.hdfs.TestFileConcurrentReader.testUnfinishedBlockCRCErrorTransferToVerySmallWrite(TestFileConcurrentReader.java:274)
Caused by: java.io.FileNotFoundException: /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/classes/hdfs-default.xml
(Too many open files)
	at java.io.FileInputStream.open(Native Method)
	at java.io.FileInputStream.<init>(FileInputStream.java:106)
	at java.io.FileInputStream.<init>(FileInputStream.java:66)
	at sun.net.www.protocol.file.FileURLConnection.connect(FileURLConnection.java:70)
	at sun.net.www.protocol.file.FileURLConnection.getInputStream(FileURLConnection.java:161)
	at com.sun.org.apache.xerces.internal.impl.XMLEntityManager.setupCurrentEntity(XMLEntityManager.java:653)
	at com.sun.org.apache.xerces.internal.impl.XMLVersionDetector.determineDocVersion(XMLVersionDetector.java:186)
	at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:771)
	at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:737)
	at com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:107)
	at com.sun.org.apache.xerces.internal.parsers.DOMParser.parse(DOMParser.java:225)
	at com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderImpl.parse(DocumentBuilderImpl.java:283)
	at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:180)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1437)


FAILED:  org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0

Error Message:
Mismatched number of datanodes

Stack Trace:
junit.framework.AssertionFailedError: Mismatched number of datanodes
	at org.apache.hadoop.hdfs.server.balancer.Balancer.chooseNodes(Balancer.java:1069)
	at org.apache.hadoop.hdfs.server.balancer.Balancer.run(Balancer.java:1511)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:247)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.test(TestBalancer.java:234)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:307)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.__CLR3_0_29j3j5brsl(TestBalancer.java:327)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0(TestBalancer.java:324)


FAILED:  org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer2

Error Message:
Mismatched number of datanodes

Stack Trace:
junit.framework.AssertionFailedError: Mismatched number of datanodes
	at org.apache.hadoop.hdfs.server.balancer.Balancer.chooseNodes(Balancer.java:1069)
	at org.apache.hadoop.hdfs.server.balancer.Balancer.run(Balancer.java:1511)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancerDefaultConstructor(TestBalancer.java:279)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancerDefaultConstructor(TestBalancer.java:376)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.__CLR3_0_2g13gq9rsu(TestBalancer.java:344)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer2(TestBalancer.java:341)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestBlockTokenWithDFS.testEnd2End

Error Message:
Mismatched number of datanodes

Stack Trace:
junit.framework.AssertionFailedError: Mismatched number of datanodes
	at org.apache.hadoop.hdfs.server.balancer.Balancer.chooseNodes(Balancer.java:1069)
	at org.apache.hadoop.hdfs.server.balancer.Balancer.run(Balancer.java:1511)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:247)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.test(TestBalancer.java:234)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:307)
	at org.apache.hadoop.hdfs.server.balancer.TestBalancer.integrationTest(TestBalancer.java:319)
	at org.apache.hadoop.hdfs.server.namenode.TestBlockTokenWithDFS.__CLR3_0_2wspf0nr53(TestBlockTokenWithDFS.java:529)
	at org.apache.hadoop.hdfs.server.namenode.TestBlockTokenWithDFS.testEnd2End(TestBlockTokenWithDFS.java:526)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestStorageRestore.testStorageRestore

Error Message:
Image file /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/secondary/current/fsimage
is corrupt with MD5 checksum of 881927486e0bcbbabc709cc7fe4a41f0 but expecting bbcc683027935545878d892b4b7981cf

Stack Trace:
java.io.IOException: Image file /grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/secondary/current/fsimage
is corrupt with MD5 checksum of 881927486e0bcbbabc709cc7fe4a41f0 but expecting bbcc683027935545878d892b4b7981cf
	at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:1062)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.doMerge(SecondaryNameNode.java:702)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.access$500(SecondaryNameNode.java:600)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doMerge(SecondaryNameNode.java:477)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:438)
	at org.apache.hadoop.hdfs.server.namenode.TestStorageRestore.__CLR3_0_2dn2tm4tdc(TestStorageRestore.java:316)
	at org.apache.hadoop.hdfs.server.namenode.TestStorageRestore.testStorageRestore(TestStorageRestore.java:286)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_2lttiju10wk(TestBlockRecovery.java:165)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedReplicas(TestBlockRecovery.java:153)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedRbwReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_2c2lg1h10x2(TestBlockRecovery.java:204)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedRbwReplicas(TestBlockRecovery.java:190)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedRwrReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_29tewcb10xl(TestBlockRecovery.java:243)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testFinalizedRwrReplicas(TestBlockRecovery.java:229)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRBWReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_2cqk51310y4(TestBlockRecovery.java:281)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRBWReplicas(TestBlockRecovery.java:269)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRBW_RWRReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_2396azp10yh(TestBlockRecovery.java:305)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRBW_RWRReplicas(TestBlockRecovery.java:293)


FAILED:  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRWRReplicas

Error Message:
null

Stack Trace:
java.lang.NullPointerException
	at org.apache.hadoop.hdfs.server.datanode.DataNode.syncBlock(DataNode.java:1883)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testSyncReplicas(TestBlockRecovery.java:144)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.__CLR3_0_2ahdlbx10yt(TestBlockRecovery.java:329)
	at org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery.testRWRReplicas(TestBlockRecovery.java:317)




Mime
View raw message