hadoop-hdfs-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Hudson Server <hud...@hudson.apache.org>
Subject Build failed in Hudson: Hadoop-Hdfs-trunk #491
Date Thu, 18 Nov 2010 16:31:26 GMT
See <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/491/changes>

Changes:

[eli] HDFS-1487. FSDirectory.removeBlock() should update diskspace count of the block owner
node. Contributed by Zhong Wang.

[eli] HDFS-1507. TestAbandonBlock should abandon a block. Contributed by Eli Collins

[eli] HDFS-259. Remove intentionally corrupt 0.13 directory layout creation. Contributed by
Todd Lipcon

[omalley] Branching for 0.22

------------------------------------------
[...truncated 756916 lines...]
    [junit] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    [junit] 	at java.lang.reflect.Method.invoke(Method.java:597)
    [junit] 	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    [junit] 	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    [junit] 	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    [junit] 	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    [junit] 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    [junit] 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    [junit] 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    [junit] 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    [junit] 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    [junit] 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    [junit] 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    [junit] 	at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    [junit] 	at junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768)
    [junit] 2010-11-18 16:32:45,327 INFO  datanode.DataNode (DataNode.java:initDataXceiver(467))
- Opened info server at 41536
    [junit] 2010-11-18 16:32:45,327 INFO  datanode.DataNode (DataXceiverServer.java:<init>(77))
- Balancing bandwith is 1048576 bytes/s
    [junit] 2010-11-18 16:32:45,329 INFO  common.Storage (DataStorage.java:recoverTransitionRead(127))
- Storage directory <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3>
is not formatted.
    [junit] 2010-11-18 16:32:45,329 INFO  common.Storage (DataStorage.java:recoverTransitionRead(128))
- Formatting ...
    [junit] 2010-11-18 16:32:45,332 INFO  common.Storage (DataStorage.java:recoverTransitionRead(127))
- Storage directory <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data4>
is not formatted.
    [junit] 2010-11-18 16:32:45,332 INFO  common.Storage (DataStorage.java:recoverTransitionRead(128))
- Formatting ...
    [junit] 2010-11-18 16:32:45,381 INFO  datanode.DataNode (FSDataset.java:registerMBean(1772))
- Registered FSDatasetStatusMBean
    [junit] 2010-11-18 16:32:45,382 INFO  datanode.DirectoryScanner (DirectoryScanner.java:<init>(149))
- scan starts at 1290109618382 with interval 21600000
    [junit] 2010-11-18 16:32:45,383 INFO  http.HttpServer (HttpServer.java:addGlobalFilter(409))
- Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
    [junit] 2010-11-18 16:32:45,384 DEBUG datanode.DataNode (DataNode.java:startInfoServer(336))
- Datanode listening on localhost:0
    [junit] 2010-11-18 16:32:45,384 INFO  http.HttpServer (HttpServer.java:start(579)) - Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener
on 0
    [junit] 2010-11-18 16:32:45,384 INFO  http.HttpServer (HttpServer.java:start(584)) - listener.getLocalPort()
returned 50189 webServer.getConnectors()[0].getLocalPort() returned 50189
    [junit] 2010-11-18 16:32:45,385 INFO  http.HttpServer (HttpServer.java:start(617)) - Jetty
bound to port 50189
    [junit] 2010-11-18 16:32:45,385 INFO  mortbay.log (?:invoke(?)) - jetty-6.1.14
    [junit] 2010-11-18 16:32:45,461 INFO  mortbay.log (?:invoke(?)) - Started SelectChannelConnector@localhost:50189
    [junit] 2010-11-18 16:32:45,461 INFO  jvm.JvmMetrics (JvmMetrics.java:init(71)) - Cannot
initialize JVM Metrics with processName=DataNode, sessionId=null - already initialized
    [junit] 2010-11-18 16:32:45,462 INFO  ipc.Server (Server.java:run(338)) - Starting SocketReader
    [junit] 2010-11-18 16:32:45,462 INFO  metrics.RpcMetrics (RpcMetrics.java:<init>(63))
- Initializing RPC Metrics with hostName=DataNode, port=43815
    [junit] 2010-11-18 16:32:45,464 INFO  metrics.RpcDetailedMetrics (RpcDetailedMetrics.java:<init>(57))
- Initializing RPC Metrics with hostName=DataNode, port=43815
    [junit] 2010-11-18 16:32:45,464 INFO  datanode.DataNode (DataNode.java:initIpcServer(427))
- dnRegistration = DatanodeRegistration(h8.grid.sp2.yahoo.net:41536, storageID=, infoPort=50189,
ipcPort=43815)
    [junit] 2010-11-18 16:32:45,466 INFO  hdfs.StateChange (FSNamesystem.java:registerDatanode(2508))
- BLOCK* NameSystem.registerDatanode: node registration from 127.0.0.1:41536 storage DS-820573405-127.0.1.1-41536-1290097965465
    [junit] 2010-11-18 16:32:45,466 INFO  net.NetworkTopology (NetworkTopology.java:add(331))
- Adding a new node: /default-rack/127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,471 INFO  datanode.DataNode (DataNode.java:register(697))
- New storage id DS-820573405-127.0.1.1-41536-1290097965465 is assigned to data-node 127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,472 INFO  datanode.DataNode (DataNode.java:run(1419)) - DatanodeRegistration(127.0.0.1:41536,
storageID=DS-820573405-127.0.1.1-41536-1290097965465, infoPort=50189, ipcPort=43815)In DataNode.run,
data = FSDataset{dirpath='<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/finalized,/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data4/current/finalized'}>
    [junit] Starting DataNode 2 with dfs.datanode.data.dir: <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/,file>:<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data6/>
    [junit] 2010-11-18 16:32:45,479 INFO  ipc.Server (Server.java:run(608)) - IPC Server Responder:
starting
    [junit] 2010-11-18 16:32:45,484 INFO  datanode.DataNode (DataNode.java:offerService(887))
- using BLOCKREPORT_INTERVAL of 21600000msec Initial delay: 0msec
    [junit] 2010-11-18 16:32:45,483 INFO  ipc.Server (Server.java:run(443)) - IPC Server listener
on 43815: starting
    [junit] 2010-11-18 16:32:45,484 INFO  ipc.Server (Server.java:run(1369)) - IPC Server
handler 0 on 43815: starting
    [junit] 2010-11-18 16:32:45,512 INFO  datanode.DataNode (DataNode.java:blockReport(1126))
- BlockReport of 0 blocks got processed in 16 msecs
    [junit] 2010-11-18 16:32:45,513 INFO  datanode.DataNode (DataNode.java:offerService(929))
- Starting Periodic block scanner.
    [junit] 2010-11-18 16:32:45,536 WARN  datanode.DataNode (DataNode.java:registerMXBean(530))
- Failed to register NameNode MXBean
    [junit] javax.management.InstanceAlreadyExistsException: HadoopInfo:type=DataNodeInfo
    [junit] 	at com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:453)
    [junit] 	at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.internal_addObject(DefaultMBeanServerInterceptor.java:1484)
    [junit] 	at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:963)
    [junit] 	at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:917)
    [junit] 	at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:312)
    [junit] 	at com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:482)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.registerMXBean(DataNode.java:528)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:498)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:281)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:266)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1556)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1499)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1466)
    [junit] 	at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:614)
    [junit] 	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:448)
    [junit] 	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:176)
    [junit] 	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:71)
    [junit] 	at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:168)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.writeSeveralPackets(TestFiDataTransferProtocol2.java:91)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.runTest17_19(TestFiDataTransferProtocol2.java:138)
    [junit] 	at org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.pipeline_Fi_19(TestFiDataTransferProtocol2.java:198)
    [junit] 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    [junit] 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    [junit] 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    [junit] 	at java.lang.reflect.Method.invoke(Method.java:597)
    [junit] 	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    [junit] 	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    [junit] 	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    [junit] 	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    [junit] 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    [junit] 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    [junit] 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    [junit] 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    [junit] 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    [junit] 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    [junit] 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    [junit] 	at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    [junit] 	at junit.framework.JUnit4TestAdapter.run(JUnit4TestAdapter.java:39)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:420)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:911)
    [junit] 	at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:768)
    [junit] 2010-11-18 16:32:45,538 INFO  datanode.DataNode (DataNode.java:initDataXceiver(467))
- Opened info server at 43306
    [junit] 2010-11-18 16:32:45,539 INFO  datanode.DataNode (DataXceiverServer.java:<init>(77))
- Balancing bandwith is 1048576 bytes/s
    [junit] 2010-11-18 16:32:45,541 INFO  common.Storage (DataStorage.java:recoverTransitionRead(127))
- Storage directory <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5>
is not formatted.
    [junit] 2010-11-18 16:32:45,541 INFO  common.Storage (DataStorage.java:recoverTransitionRead(128))
- Formatting ...
    [junit] 2010-11-18 16:32:45,543 INFO  common.Storage (DataStorage.java:recoverTransitionRead(127))
- Storage directory <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data6>
is not formatted.
    [junit] 2010-11-18 16:32:45,544 INFO  common.Storage (DataStorage.java:recoverTransitionRead(128))
- Formatting ...
    [junit] 2010-11-18 16:32:45,582 INFO  datanode.DataNode (FSDataset.java:registerMBean(1772))
- Registered FSDatasetStatusMBean
    [junit] 2010-11-18 16:32:45,582 INFO  datanode.DirectoryScanner (DirectoryScanner.java:<init>(149))
- scan starts at 1290115083582 with interval 21600000
    [junit] 2010-11-18 16:32:45,584 INFO  http.HttpServer (HttpServer.java:addGlobalFilter(409))
- Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
    [junit] 2010-11-18 16:32:45,584 DEBUG datanode.DataNode (DataNode.java:startInfoServer(336))
- Datanode listening on localhost:0
    [junit] 2010-11-18 16:32:45,585 INFO  http.HttpServer (HttpServer.java:start(579)) - Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener
on 0
    [junit] 2010-11-18 16:32:45,585 INFO  http.HttpServer (HttpServer.java:start(584)) - listener.getLocalPort()
returned 60767 webServer.getConnectors()[0].getLocalPort() returned 60767
    [junit] 2010-11-18 16:32:45,586 INFO  http.HttpServer (HttpServer.java:start(617)) - Jetty
bound to port 60767
    [junit] 2010-11-18 16:32:45,586 INFO  mortbay.log (?:invoke(?)) - jetty-6.1.14
    [junit] 2010-11-18 16:32:45,675 INFO  mortbay.log (?:invoke(?)) - Started SelectChannelConnector@localhost:60767
    [junit] 2010-11-18 16:32:45,676 INFO  jvm.JvmMetrics (JvmMetrics.java:init(71)) - Cannot
initialize JVM Metrics with processName=DataNode, sessionId=null - already initialized
    [junit] 2010-11-18 16:32:45,677 INFO  ipc.Server (Server.java:run(338)) - Starting SocketReader
    [junit] 2010-11-18 16:32:45,677 INFO  metrics.RpcMetrics (RpcMetrics.java:<init>(63))
- Initializing RPC Metrics with hostName=DataNode, port=52910
    [junit] 2010-11-18 16:32:45,678 INFO  metrics.RpcDetailedMetrics (RpcDetailedMetrics.java:<init>(57))
- Initializing RPC Metrics with hostName=DataNode, port=52910
    [junit] 2010-11-18 16:32:45,678 INFO  datanode.DataNode (DataNode.java:initIpcServer(427))
- dnRegistration = DatanodeRegistration(h8.grid.sp2.yahoo.net:43306, storageID=, infoPort=60767,
ipcPort=52910)
    [junit] 2010-11-18 16:32:45,680 INFO  hdfs.StateChange (FSNamesystem.java:registerDatanode(2508))
- BLOCK* NameSystem.registerDatanode: node registration from 127.0.0.1:43306 storage DS-835262486-127.0.1.1-43306-1290097965679
    [junit] 2010-11-18 16:32:45,680 INFO  net.NetworkTopology (NetworkTopology.java:add(331))
- Adding a new node: /default-rack/127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,684 INFO  datanode.DataNode (DataNode.java:register(697))
- New storage id DS-835262486-127.0.1.1-43306-1290097965679 is assigned to data-node 127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,685 INFO  datanode.DataNode (DataNode.java:run(1419)) - DatanodeRegistration(127.0.0.1:43306,
storageID=DS-835262486-127.0.1.1-43306-1290097965679, infoPort=60767, ipcPort=52910)In DataNode.run,
data = FSDataset{dirpath='<https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/finalized,/grid/0/hudson/hudson-slave/workspace/Hadoop-Hdfs-trunk/trunk/build-fi/test/data/dfs/data/data6/current/finalized'}>
    [junit] 2010-11-18 16:32:45,685 INFO  ipc.Server (Server.java:run(608)) - IPC Server Responder:
starting
    [junit] 2010-11-18 16:32:45,686 INFO  ipc.Server (Server.java:run(443)) - IPC Server listener
on 52910: starting
    [junit] 2010-11-18 16:32:45,686 INFO  ipc.Server (Server.java:run(1369)) - IPC Server
handler 0 on 52910: starting
    [junit] 2010-11-18 16:32:45,687 INFO  datanode.DataNode (DataNode.java:offerService(887))
- using BLOCKREPORT_INTERVAL of 21600000msec Initial delay: 0msec
    [junit] 2010-11-18 16:32:45,692 INFO  datanode.DataNode (DataNode.java:blockReport(1126))
- BlockReport of 0 blocks got processed in 2 msecs
    [junit] 2010-11-18 16:32:45,692 INFO  datanode.DataNode (DataNode.java:offerService(929))
- Starting Periodic block scanner.
    [junit] 2010-11-18 16:32:45,695 DEBUG hdfs.DFSClient (DFSClient.java:create(629)) - /pipeline_Fi_19/foo:
masked=rwxr-xr-x
    [junit] 2010-11-18 16:32:45,695 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,708 INFO  FSNamesystem.audit (FSNamesystem.java:logAuditEvent(148))
- ugi=hudson	ip=/127.0.0.1	cmd=create	src=/pipeline_Fi_19/foo	dst=null	perm=hudson:supergroup:rw-r--r--
    [junit] 2010-11-18 16:32:45,710 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=0, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=0
    [junit] 2010-11-18 16:32:45,711 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=0, src=/pipeline_Fi_19/foo, bytesCurBlock=1024, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,711 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 0
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(444)) -
Allocating new block
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,712 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=1, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=1024
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=1, src=/pipeline_Fi_19/foo, bytesCurBlock=2048, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 1
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,713 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=2, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=2048
    [junit] 2010-11-18 16:32:45,713 INFO  hdfs.StateChange (FSNamesystem.java:allocateBlock(1753))
- BLOCK* NameSystem.allocateBlock: /pipeline_Fi_19/foo. blk_-1087155876419230760_1001{blockUCState=UNDER_CONSTRUCTION,
primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[127.0.0.1:41536|RBW], ReplicaUnderConstruction[127.0.0.1:43306|RBW],
ReplicaUnderConstruction[127.0.0.1:36464|RBW]]}
    [junit] 2010-11-18 16:32:45,714 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=2, src=/pipeline_Fi_19/foo, bytesCurBlock=3072, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,714 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 2
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,715 INFO  protocol.ClientProtocolAspects (ClientProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_protocol_ClientProtocolAspects$1$7076326d(35))
- FI: addBlock Pipeline[127.0.0.1:41536, 127.0.0.1:43306, 127.0.0.1:36464]
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient (DFSOutputStream.java:createBlockOutputStream(881))
- pipeline = 127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient (DFSOutputStream.java:createBlockOutputStream(881))
- pipeline = 127.0.0.1:43306
    [junit] 2010-11-18 16:32:45,715 DEBUG hdfs.DFSClient (DFSOutputStream.java:createBlockOutputStream(881))
- pipeline = 127.0.0.1:36464
    [junit] 2010-11-18 16:32:45,716 DEBUG hdfs.DFSClient (DFSOutputStream.java:createBlockOutputStream(891))
- Connecting to 127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,716 DEBUG datanode.DataNode (DataXceiver.java:<init>(86))
- Number of active connections is: 1
    [junit] 2010-11-18 16:32:45,716 DEBUG hdfs.DFSClient (DFSOutputStream.java:createBlockOutputStream(900))
- Send buf size 131071
    [junit] 2010-11-18 16:32:45,717 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,717 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=3, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=3072
    [junit] 2010-11-18 16:32:45,717 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:45,717 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=3, src=/pipeline_Fi_19/foo, bytesCurBlock=4096, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil (FiTestUtil.java:initialValue(37))
- Thread[DataXceiver for client /127.0.0.1:44424 [Waiting for operation],5,dataXceiverServer]:
seed=-5326246243198308466
    [junit] 2010-11-18 16:32:45,718 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Waiting for operation] sleeps for 1556ms
    [junit] 2010-11-18 16:32:45,718 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 3
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=4, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=4096
    [junit] 2010-11-18 16:32:45,719 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=4, src=/pipeline_Fi_19/foo, bytesCurBlock=5120, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 4
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,720 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=5, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=5120
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1221))
- DFSClient writeChunk packet full seqno=5, src=/pipeline_Fi_19/foo, bytesCurBlock=6144, blockSize=1048576,
appendChunk=false
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 5
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient (DFSOutputStream.java:computePacketChunkSize(1144))
- computePacketChunkSize: src=/pipeline_Fi_19/foo, chunkSize=516, chunksPerPacket=2, packetSize=1057
    [junit] 2010-11-18 16:32:45,721 DEBUG hdfs.DFSClient (DFSOutputStream.java:writeChunk(1202))
- DFSClient writeChunk allocating new packet seqno=6, src=/pipeline_Fi_19/foo, packetSize=1057,
chunksPerPacket=2, bytesCurBlock=6144
    [junit] 2010-11-18 16:32:45,722 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 6
    [junit] 2010-11-18 16:32:45,722 INFO  hdfs.DFSClientAspects (DFSClientAspects.aj:ajc$before$org_apache_hadoop_hdfs_DFSClientAspects$5$5ba7280d(86))
- FI: before pipelineClose:
    [junit] 2010-11-18 16:32:45,722 DEBUG hdfs.DFSClient (DFSOutputStream.java:queueCurrentPacket(1157))
- Queued packet 7
    [junit] 2010-11-18 16:32:45,723 DEBUG hdfs.DFSClient (DFSOutputStream.java:waitForAckedSeqno(1408))
- Waiting for ack for: 7
    [junit] 2010-11-18 16:32:47,275 DEBUG datanode.DataNode (DataXceiver.java:opWriteBlock(246))
- writeBlock receive buf size 131071 tcp no delay true
    [junit] 2010-11-18 16:32:47,275 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(251))
- Receiving block blk_-1087155876419230760_1001 src: /127.0.0.1:44424 dest: /127.0.0.1:41536
    [junit] 2010-11-18 16:32:47,276 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(176))
- writeTo blockfile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/rbw/blk_-1087155876419230760>
of size 0
    [junit] 2010-11-18 16:32:47,276 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(178))
- writeTo metafile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data3/current/rbw/blk_-1087155876419230760_1001.meta>
of size 0
    [junit] 2010-11-18 16:32:47,277 DEBUG datanode.DataNode (DataXceiver.java:<init>(86))
- Number of active connections is: 1
    [junit] 2010-11-18 16:32:47,287 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,287 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=1, duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil (FiTestUtil.java:initialValue(37))
- Thread[DataXceiver for client /127.0.0.1:60263 [Waiting for operation],5,dataXceiverServer]:
seed=7365113689811878303
    [junit] 2010-11-18 16:32:47,287 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:60263 [Waiting for operation] sleeps for 506ms
    [junit] 2010-11-18 16:32:47,794 DEBUG datanode.DataNode (DataXceiver.java:opWriteBlock(246))
- writeBlock receive buf size 131071 tcp no delay true
    [junit] 2010-11-18 16:32:47,794 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(251))
- Receiving block blk_-1087155876419230760_1001 src: /127.0.0.1:60263 dest: /127.0.0.1:43306
    [junit] 2010-11-18 16:32:47,795 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(176))
- writeTo blockfile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/rbw/blk_-1087155876419230760>
of size 0
    [junit] 2010-11-18 16:32:47,795 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(178))
- writeTo metafile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data5/current/rbw/blk_-1087155876419230760_1001.meta>
of size 0
    [junit] 2010-11-18 16:32:47,796 DEBUG datanode.DataNode (DataXceiver.java:<init>(86))
- Number of active connections is: 1
    [junit] 2010-11-18 16:32:47,796 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$1$8f59fdd7(51))
- FI: receiverOp WRITE_BLOCK, datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:47,796 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$3$3251489(73))
- FI: receiverOpWriteBlock
    [junit] 2010-11-18 16:32:47,796 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=2, duration=[0, 3000), datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:47,797 INFO  fi.FiTestUtil (FiTestUtil.java:initialValue(37))
- Thread[DataXceiver for client /127.0.0.1:59520 [Waiting for operation],5,dataXceiverServer]:
seed=-3002906034623524893
    [junit] 2010-11-18 16:32:47,797 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:59520 [Waiting for operation] sleeps for 1956ms
    [junit] 2010-11-18 16:32:49,753 DEBUG datanode.DataNode (DataXceiver.java:opWriteBlock(246))
- writeBlock receive buf size 131071 tcp no delay true
    [junit] 2010-11-18 16:32:49,753 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(251))
- Receiving block blk_-1087155876419230760_1001 src: /127.0.0.1:59520 dest: /127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,754 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(176))
- writeTo blockfile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data1/current/rbw/blk_-1087155876419230760>
of size 0
    [junit] 2010-11-18 16:32:49,754 DEBUG datanode.DataNode (ReplicaInPipeline.java:createStreams(178))
- writeTo metafile is <https://hudson.apache.org/hudson/job/Hadoop-Hdfs-trunk/ws/trunk/build-fi/test/data/dfs/data/data1/current/rbw/blk_-1087155876419230760_1001.meta>
of size 0
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(371))
- Datanode 0 forwarding connect ack to upstream firstbadlink is 
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
- FI: statusRead SUCCESS, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:49,755 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,755 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=1, duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:49,756 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:60263 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 1646ms
    [junit] 2010-11-18 16:32:49,755 DEBUG datanode.DataNode (BlockReceiver.java:run(843))
- PacketResponder 0 seqno = -2 for block blk_-1087155876419230760_1001 waiting for local datanode
to finish write.
    [junit] 2010-11-18 16:32:49,755 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=2, duration=[0, 3000), datanode=127.0.0.1:36464
    [junit] 2010-11-18 16:32:49,756 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:59520 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 1012ms
    [junit] 2010-11-18 16:32:51,402 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(338))
- Datanode 1 got response for connect ack  from downstream datanode with firstbadlink as 
    [junit] 2010-11-18 16:32:51,402 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(371))
- Datanode 1 forwarding connect ack to upstream firstbadlink is 
    [junit] 2010-11-18 16:32:51,403 INFO  datanode.DataTransferProtocolAspects (DataTransferProtocolAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_server_datanode_DataTransferProtocolAspects$2$d4f6605f(61))
- FI: statusRead SUCCESS, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:51,403 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=1, duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 1803ms
    [junit] 2010-11-18 16:32:51,403 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:60263 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 2647ms
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(338))
- Datanode 2 got response for connect ack  from downstream datanode with firstbadlink as 
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.DataNode (DataXceiver.java:opWriteBlock(371))
- Datanode 2 forwarding connect ack to upstream firstbadlink is 
    [junit] 2010-11-18 16:32:53,207 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:53,207 INFO  hdfs.DFSClientAspects (DFSClientAspects.aj:ajc$afterReturning$org_apache_hadoop_hdfs_DFSClientAspects$2$9396d2df(48))
- FI: after pipelineInitNonAppend: hasError=false errorIndex=-1
    [junit] 2010-11-18 16:32:53,207 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:53,208 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 2638ms
    [junit] 2010-11-18 16:32:53,208 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:0 offsetInBlock:0
lastPacketInBlock:false lastByteOffsetInBlock: 1024
    [junit] 2010-11-18 16:32:53,208 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:1 offsetInBlock:1024
lastPacketInBlock:false lastByteOffsetInBlock: 2048
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:2 offsetInBlock:2048
lastPacketInBlock:false lastByteOffsetInBlock: 3072
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:3 offsetInBlock:3072
lastPacketInBlock:false lastByteOffsetInBlock: 4096
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:4 offsetInBlock:4096
lastPacketInBlock:false lastByteOffsetInBlock: 5120
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:5 offsetInBlock:5120
lastPacketInBlock:false lastByteOffsetInBlock: 6144
    [junit] 2010-11-18 16:32:53,209 DEBUG hdfs.DFSClient (DFSOutputStream.java:run(496)) -
DataStreamer block blk_-1087155876419230760_1001 sending packet packet seqno:6 offsetInBlock:6144
lastPacketInBlock:false lastByteOffsetInBlock: 6170
    [junit] 2010-11-18 16:32:55,846 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:55,847 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:55,847 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 863ms
    [junit] 2010-11-18 16:32:56,710 DEBUG datanode.DataNode (BlockReceiver.java:receivePacket(456))
- Receiving one packet for block blk_-1087155876419230760_1001 of length 1024 seqno 0 offsetInBlock
0 lastPacketInBlock false
    [junit] 2010-11-18 16:32:56,710 DEBUG datanode.DataNode (BlockReceiver.java:enqueue(788))
- PacketResponder 2 adding seqno 0 to ack queue.
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
- FI: callWritePacketToDisk
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:56,710 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=1, duration=[0, 3000), datanode=127.0.0.1:43306
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:60263 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 2454ms
    [junit] 2010-11-18 16:32:56,711 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 353ms
    [junit] 2010-11-18 16:32:57,065 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,065 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,065 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 377ms
    [junit] 2010-11-18 16:32:57,442 DEBUG datanode.DataNode (BlockReceiver.java:receivePacket(456))
- Receiving one packet for block blk_-1087155876419230760_1001 of length 1024 seqno 1 offsetInBlock
1024 lastPacketInBlock false
    [junit] 2010-11-18 16:32:57,442 DEBUG datanode.DataNode (BlockReceiver.java:enqueue(788))
- PacketResponder 2 adding seqno 1 to ack queue.
    [junit] 2010-11-18 16:32:57,443 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$2$56c32214(71))
- FI: callWritePacketToDisk
    [junit] 2010-11-18 16:32:57,443 INFO  datanode.BlockReceiverAspects (BlockReceiverAspects.aj:ajc$before$org_apache_hadoop_hdfs_server_datanode_BlockReceiverAspects$1$4c211928(53))
- FI: callReceivePacket, datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,443 INFO  fi.FiTestUtil (DataTransferTestUtil.java:run(344))
- FI: SleepAction:pipeline_Fi_19, index=0, duration=[0, 3000), datanode=127.0.0.1:41536
    [junit] 2010-11-18 16:32:57,443 INFO  fi.FiTestUtil (FiTestUtil.java:sleep(92)) - DataXceiver
for client /127.0.0.1:44424 [Receiving block blk_-1087155876419230760_1001 client=DFSClient_-244463401]
sleeps for 1915ms
Build timed out. Aborting
Publishing Javadoc
Archiving artifacts
Recording test results
Recording fingerprints
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure


Mime
View raw message