hadoop-mapreduce-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Hadoop-Mapreduce-22-branch - Build # 87 - Failure
Date Thu, 03 Nov 2011 00:36:47 GMT
See https://builds.apache.org/job/Hadoop-Mapreduce-22-branch/87/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 500297 lines...]
    [junit] 11/11/03 00:32:26 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread.
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:48938,
storageID=DS-475485639-67.195.138.25-48938-1320280345227, infoPort=36358, ipcPort=59430):Finishing
DataNode in: FSDataset{dirpath='/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build/contrib/raid/test/data/dfs/data/data3/current/finalized,/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build/contrib/raid/test/data/dfs/data/data4/current/finalized'}
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping server on 59430
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: Waiting for threadgroup to exit, active
threads is 0
    [junit] 11/11/03 00:32:26 INFO datanode.FSDatasetAsyncDiskService: Shutting down all async
disk service threads...
    [junit] 11/11/03 00:32:26 INFO datanode.FSDatasetAsyncDiskService: All async disk service
threads have been shut down.
    [junit] 11/11/03 00:32:26 WARN datanode.FSDatasetAsyncDiskService: AsyncDiskService has
already shut down.
    [junit] 11/11/03 00:32:26 INFO hdfs.MiniDFSCluster: Shutting down DataNode 0
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping server on 42200
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 0 on 42200: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 2 on 42200: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 1 on 42200: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping IPC Server listener on 42200
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping IPC Server Responder
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: Waiting for threadgroup to exit, active
threads is 25
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: Waiting for threadgroup to exit, active
threads is 0
    [junit] 11/11/03 00:32:26 INFO datanode.DataBlockScanner: Exiting DataBlockScanner thread.
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: DatanodeRegistration(127.0.0.1:44434,
storageID=DS-908436179-67.195.138.25-44434-1320280345099, infoPort=55557, ipcPort=42200):Finishing
DataNode in: FSDataset{dirpath='/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build/contrib/raid/test/data/dfs/data/data1/current/finalized,/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build/contrib/raid/test/data/dfs/data/data2/current/finalized'}
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping server on 42200
    [junit] 11/11/03 00:32:26 INFO datanode.DataNode: Waiting for threadgroup to exit, active
threads is 0
    [junit] 11/11/03 00:32:26 INFO datanode.FSDatasetAsyncDiskService: Shutting down all async
disk service threads...
    [junit] 11/11/03 00:32:26 INFO datanode.FSDatasetAsyncDiskService: All async disk service
threads have been shut down.
    [junit] 11/11/03 00:32:26 WARN datanode.FSDatasetAsyncDiskService: AsyncDiskService has
already shut down.
    [junit] 11/11/03 00:32:26 WARN namenode.FSNamesystem: ReplicationMonitor thread received
InterruptedException.java.lang.InterruptedException: sleep interrupted
    [junit] 11/11/03 00:32:26 WARN namenode.DecommissionManager: Monitor interrupted: java.lang.InterruptedException:
sleep interrupted
    [junit] 11/11/03 00:32:26 INFO namenode.FSEditLog: Number of transactions: 14 Total time
for transactions(ms): 2Number of transactions batched in Syncs: 0 Number of syncs: 7 SyncTimes(ms):
5 2 
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping server on 58221
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 0 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 2 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 5 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 8 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 9 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 1 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping IPC Server listener on 58221
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 4 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 7 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 6 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: IPC Server handler 3 on 58221: exiting
    [junit] 11/11/03 00:32:26 INFO ipc.Server: Stopping IPC Server Responder
    [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.89 sec

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build.xml:817: The
following error occurred while executing this line:
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/build.xml:796: The
following error occurred while executing this line:
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/src/contrib/build.xml:87:
The following error occurred while executing this line:
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-22-branch/trunk/src/contrib/raid/build.xml:60:
Tests failed!

Total time: 193 minutes 53 seconds
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Recording test results
Publishing Javadoc
Recording fingerprints
Updating MAPREDUCE-3139
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
4 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.hadoop.mapred.TestFairSchedulerSystem

Error Message:
java.net.ConnectException: Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException:
Connection refused

Stack Trace:
java.lang.RuntimeException: java.net.ConnectException: Call to localhost/127.0.0.1:0 failed
on connection exception: java.net.ConnectException: Connection refused
	at org.apache.hadoop.mapred.MiniMRCluster.waitUntilIdle(MiniMRCluster.java:336)
	at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:546)
	at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:483)
	at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:475)
	at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:418)
	at org.apache.hadoop.mapred.TestFairSchedulerSystem.setUp(TestFairSchedulerSystem.java:74)
Caused by: java.net.ConnectException: Call to localhost/127.0.0.1:0 failed on connection exception:
java.net.ConnectException: Connection refused
	at org.apache.hadoop.ipc.Client.wrapException(Client.java:1055)
	at org.apache.hadoop.ipc.Client.call(Client.java:1031)
	at org.apache.hadoop.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:198)
	at $Proxy6.getProtocolVersion(Unknown Source)
	at org.apache.hadoop.ipc.WritableRpcEngine.getProxy(WritableRpcEngine.java:235)
	at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:275)
	at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:249)
	at org.apache.hadoop.mapreduce.Cluster.createRPCProxy(Cluster.java:86)
	at org.apache.hadoop.mapreduce.Cluster.createClient(Cluster.java:98)
	at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:74)
	at org.apache.hadoop.mapred.JobClient.init(JobClient.java:456)
	at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:435)
	at org.apache.hadoop.mapred.MiniMRCluster.waitUntilIdle(MiniMRCluster.java:322)
Caused by: java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
	at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:373)
	at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:416)
	at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:504)
	at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:206)
	at org.apache.hadoop.ipc.Client.getConnection(Client.java:1164)
	at org.apache.hadoop.ipc.Client.call(Client.java:1008)


FAILED:  org.apache.hadoop.raid.TestRaidNode.testPathFilter

Error Message:
Too many open files  at sun.nio.ch.IOUtil.initPipe(Native Method)  at sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:49)
 at sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)  at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
 at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
 at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)  at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:159)
 at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:132)  at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
 at java.io.BufferedInputStream.read1(BufferedInputStream.java:258)  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
 at java.io.DataInputStream.read(DataInputStream.java:132)  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:122)
 at org.apache.hadoop.hdfs.BlockReader.readChunk(BlockReader.java:297)  at org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:273)
 at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:225)  at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:193)
 at org.apache.hadoop.hdfs.BlockReader.read(BlockReader.java:136)  at org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:466)
 at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:517)  at java.io.DataInputStream.read(DataInputStream.java:132)
 at org.apache.hadoop.raid.ParityInputStream.readExact(ParityInputStream.java:138)  at org.apache.hadoop.raid.ParityInputStream.makeAvailable(ParityInputStream.java:117)
 at org.apache.hadoop.raid.ParityInputStream.drain(ParityInputStream.java:95)  at org.apache.hadoop.raid.XORDecoder.fixErasedBlock(XORDecoder.java:74)
 at org.apache.hadoop.raid.Decoder.decodeFile(Decoder.java:147)  at org.apache.hadoop.raid.RaidNode.unRaid(RaidNode.java:867)
 at org.apache.hadoop.raid.RaidNode.recoverFile(RaidNode.java:333)  at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown
Source)  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)  at org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:349)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1482)  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1478)
 at java.security.AccessController.doPrivileged(Native Method)  at javax.security.auth.Subject.doAs(Subject.java:396)
 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1153) 
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1476) 

Stack Trace:
java.io.IOException: Too many open files
	at sun.nio.ch.IOUtil.initPipe(Native Method)
	at sun.nio.ch.EPollSelectorImpl.<init>(EPollSelectorImpl.java:49)
	at sun.nio.ch.EPollSelectorProvider.openSelector(EPollSelectorProvider.java:18)
	at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.get(SocketIOWithTimeout.java:407)
	at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:322)
	at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:157)
	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:159)
	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:132)
	at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
	at java.io.BufferedInputStream.read1(BufferedInputStream.java:258)
	at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
	at java.io.DataInputStream.read(DataInputStream.java:132)
	at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:122)
	at org.apache.hadoop.hdfs.BlockReader.readChunk(BlockReader.java:297)
	at org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:273)
	at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:225)
	at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:193)
	at org.apache.hadoop.hdfs.BlockReader.read(BlockReader.java:136)
	at org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:466)
	at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:517)
	at java.io.DataInputStream.read(DataInputStream.java:132)
	at org.apache.hadoop.raid.ParityInputStream.readExact(ParityInputStream.java:138)
	at org.apache.hadoop.raid.ParityInputStream.makeAvailable(ParityInputStream.java:117)
	at org.apache.hadoop.raid.ParityInputStream.drain(ParityInputStream.java:95)
	at org.apache.hadoop.raid.XORDecoder.fixErasedBlock(XORDecoder.java:74)
	at org.apache.hadoop.raid.Decoder.decodeFile(Decoder.java:147)
	at org.apache.hadoop.raid.RaidNode.unRaid(RaidNode.java:867)
	at org.apache.hadoop.raid.RaidNode.recoverFile(RaidNode.java:333)
	at org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:349)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1482)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1478)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1153)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1476)

	at org.apache.hadoop.ipc.Client.call(Client.java:1028)
	at org.apache.hadoop.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:198)
	at $Proxy11.recoverFile(Unknown Source)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:84)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
	at $Proxy11.recoverFile(Unknown Source)
	at org.apache.hadoop.raid.RaidShell.recover(RaidShell.java:272)
	at org.apache.hadoop.raid.TestRaidNode.simulateError(TestRaidNode.java:576)
	at org.apache.hadoop.raid.TestRaidNode.doTestPathFilter(TestRaidNode.java:331)
	at org.apache.hadoop.raid.TestRaidNode.testPathFilter(TestRaidNode.java:257)


FAILED:  org.apache.hadoop.streaming.TestDumpTypedBytes.testDumping

Error Message:
port out of range:-1

Stack Trace:
java.lang.IllegalArgumentException: port out of range:-1
	at java.net.InetSocketAddress.<init>(InetSocketAddress.java:118)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:519)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:459)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1153)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:459)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.activate(NameNode.java:403)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:387)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:576)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:569)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1538)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:445)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:378)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:259)
	at org.apache.hadoop.streaming.TestDumpTypedBytes.testDumping(TestDumpTypedBytes.java:42)


FAILED:  org.apache.hadoop.streaming.TestLoadTypedBytes.testLoading

Error Message:
port out of range:-1

Stack Trace:
java.lang.IllegalArgumentException: port out of range:-1
	at java.net.InetSocketAddress.<init>(InetSocketAddress.java:118)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:519)
	at org.apache.hadoop.hdfs.server.namenode.NameNode$1.run(NameNode.java:459)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1153)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:459)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.activate(NameNode.java:403)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:387)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:576)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:569)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1538)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:445)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:378)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:259)
	at org.apache.hadoop.streaming.TestLoadTypedBytes.testLoading(TestLoadTypedBytes.java:42)




Mime
View raw message