hbase-builds mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: HBase-Trunk_matrix » latest1.7,yahoo-not-h2 #1175
Date Tue, 05 Jul 2016 23:03:23 GMT
See <https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/1175/changes>

Changes:

[busbey] HBASE-15985 clarify promises about edits from replication in ref guide

[apurtell] HBASE-16091 Canary takes lot more time when there are delete markers in

[eclark] HBASE-16177 In dev mode thrift server can't be run

[tedyu] HBASE-16144 Revert - test failure in

[eclark] HBASE-16087 Replication shouldn't start on a master if if only hosts

------------------------------------------
[...truncated 5609 lines...]
Running org.apache.hadoop.hbase.mapreduce.TestImportTsv
Running org.apache.hadoop.hbase.mapreduce.TestWALPlayer
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.544 sec - in org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery
Running org.apache.hadoop.hbase.mapreduce.TestImportExport
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 270.164 sec - in org.apache.hadoop.hbase.client.TestReplicasClient
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.606 sec - in org.apache.hadoop.hbase.mapreduce.TestWALPlayer
Running org.apache.hadoop.hbase.mapreduce.TestRowCounter
Running org.apache.hadoop.hbase.mapreduce.TestImportTSVWithVisibilityLabels
Running org.apache.hadoop.hbase.mapreduce.TestHashTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.546 sec - in org.apache.hadoop.hbase.mapreduce.TestHashTable
Running org.apache.hadoop.hbase.mapreduce.TestImportTSVWithTTLs
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.569 sec - in org.apache.hadoop.hbase.mapreduce.TestImportTSVWithTTLs
Running org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 75.628 sec - in org.apache.hadoop.hbase.mapreduce.TestRowCounter
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 89.639 sec - in org.apache.hadoop.hbase.mapreduce.TestImportExport
Running org.apache.hadoop.hbase.mapreduce.TestWALRecordReader
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.38 sec - in org.apache.hadoop.hbase.mapreduce.TestWALRecordReader
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.773 sec - in org.apache.hadoop.hbase.mapreduce.TestImportTSVWithVisibilityLabels
Running org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Running org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Running org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 149.115 sec - in org.apache.hadoop.hbase.mapreduce.TestImportTsv
Running org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Tests run: 16, Failures: 0, Errors: 1, Skipped: 14, Time elapsed: 90.296 sec <<<
FAILURE! - in org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2
testMRIncrementalLoadWithPutSortReducer(org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2)
 Time elapsed: 87.966 sec  <<< ERROR!
java.lang.RuntimeException: org.apache.hadoop.hbase.client.ScannerTimeoutException: 60015ms
passed since the last invocation, timeout is currently set to 60000
	at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:433)
	at org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:363)
	at org.apache.hadoop.hbase.client.ClientSimpleScanner.next(ClientSimpleScanner.java:51)
	at org.apache.hadoop.hbase.MetaTableAccessor.scanMeta(MetaTableAccessor.java:776)
	at org.apache.hadoop.hbase.MetaTableAccessor.scanMeta(MetaTableAccessor.java:702)
	at org.apache.hadoop.hbase.MetaTableAccessor.getTableRegionsAndLocations(MetaTableAccessor.java:624)
	at org.apache.hadoop.hbase.client.ConnectionImplementation.isTableAvailable(ConnectionImplementation.java:520)
	at org.apache.hadoop.hbase.client.HBaseAdmin.isTableAvailable(HBaseAdmin.java:846)
	at org.apache.hadoop.hbase.HBaseTestingUtility$11.evaluate(HBaseTestingUtility.java:3905)
	at org.apache.hadoop.hbase.Waiter.waitFor(Waiter.java:191)
	at org.apache.hadoop.hbase.Waiter.waitFor(Waiter.java:143)
	at org.apache.hadoop.hbase.HBaseTestingUtility.waitFor(HBaseTestingUtility.java:3816)
	at org.apache.hadoop.hbase.HBaseTestingUtility.waitTableAvailable(HBaseTestingUtility.java:2934)
	at org.apache.hadoop.hbase.HBaseTestingUtility.waitTableAvailable(HBaseTestingUtility.java:2917)
	at org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2.doIncrementalLoadTest(TestHFileOutputFormat2.java:650)
	at org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2.testMRIncrementalLoadWithPutSortReducer(TestHFileOutputFormat2.java:539)
Caused by: org.apache.hadoop.hbase.UnknownScannerException: org.apache.hadoop.hbase.UnknownScannerException:
Unknown scanner '29'. This can happen due to any of the following reasons: a) Scanner id given
is wrong, b) Scanner lease expired because of long wait between consecutive client checkins,
c) Server may be closing down, d) RegionServer restart during upgrade.
If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period'
configuration.
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2546)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34818)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2212)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:118)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:189)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:169)

	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
	at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
	at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
	at org.apache.hadoop.hbase.ipc.AsyncCall.setFailed(AsyncCall.java:159)
	at org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:81)
	at org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
	at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
	at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
	at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
	at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
	at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
	at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
	at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
	at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
	at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
	at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
	at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: org.apache.hadoop.hbase.UnknownScannerException:
Unknown scanner '29'. This can happen due to any of the following reasons: a) Scanner id given
is wrong, b) Scanner lease expired because of long wait between consecutive client checkins,
c) Server may be closing down, d) RegionServer restart during upgrade.
If the issue is due to reason (b), a possible fix would be increasing the value of'hbase.client.scanner.timeout.period'
configuration.
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2546)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34818)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2212)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:118)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:189)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:169)

	at org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.createRemoteException(AsyncServerResponseHandler.java:120)
	at org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:76)
	at org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
	at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
	at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
	at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
	at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
	at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
	at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
	at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
	at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
	at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
	at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
	at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
	at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
	at java.lang.Thread.run(Thread.java:745)

Running org.apache.hadoop.hbase.mapreduce.TestSyncTable
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 48.238 sec - in org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.824 sec - in org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 46.15 sec - in org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Running org.apache.hadoop.hbase.io.TestFileLink
Running org.apache.hadoop.hbase.mapreduce.TestCopyTable
Running org.apache.hadoop.hbase.TestMetaTableAccessor
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.208 sec - in org.apache.hadoop.hbase.io.TestFileLink
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.529 sec - in org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.58 sec - in org.apache.hadoop.hbase.mapreduce.TestSyncTable
Running org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Running org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.055 sec - in org.apache.hadoop.hbase.TestMetaTableAccessor
Running org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.296 sec - in org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.045 sec - in org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Running org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.007 sec - in org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Running org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.775 sec - in org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 67.646 sec - in org.apache.hadoop.hbase.mapreduce.TestCopyTable
Running org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 162.543 sec - in org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 66.022 sec - in org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.682 sec - in org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.732 sec - in org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Running org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 95.124 sec - in org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Running org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 56, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 56.517 sec - in org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.572 sec - in org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Running org.apache.hadoop.hbase.filter.TestScanRowPrefix
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.911 sec - in org.apache.hadoop.hbase.filter.TestScanRowPrefix
Running org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Running org.apache.hadoop.hbase.filter.TestFilterWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.029 sec - in org.apache.hadoop.hbase.filter.TestFilterWrapper
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 127.48 sec - in org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Running org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd
Running org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Tests run: 36, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 83.301 sec - in org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.109 sec - in org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.593 sec - in org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Running org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.591 sec - in org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Tests run: 72, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 180.893 sec - in org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.244 sec - in org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.027 sec - in org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 311.151 sec - in org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd

Results :

Failed tests: 
  TestModifyNamespaceProcedure.testRollbackAndDoubleExecution:278 expected null, but was:<bar>
Tests in error: 
  TestHFileOutputFormat2.testMRIncrementalLoadWithPutSortReducer:539->doIncrementalLoadTest:650
» Runtime

Tests run: 2069, Failures: 1, Errors: 1, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache HBase ...................................... SUCCESS [  3.409 s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [  0.874 s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [  0.287 s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [  0.232 s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [  3.255 s]
[INFO] Apache HBase - Common ............................. SUCCESS [01:42 min]
[INFO] Apache HBase - Procedure .......................... SUCCESS [01:40 min]
[INFO] Apache HBase - Client ............................. SUCCESS [ 48.830 s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [  7.992 s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [ 10.822 s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [ 11.072 s]
[INFO] Apache HBase - Server ............................. FAILURE [  01:14 h]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - RSGroup ............................ SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Spark .............................. SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] Apache HBase - Archetypes ......................... SKIPPED
[INFO] Apache HBase - Exemplar for hbase-client archetype  SKIPPED
[INFO] Apache HBase - Exemplar for hbase-shaded-client archetype  SKIPPED
[INFO] Apache HBase - Archetype builder .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:19 h
[INFO] Finished at: 2016-07-05T23:01:25+00:00
[INFO] Final Memory: 70M/508M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test
(secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/ws/hbase-server/target/surefire-reports>
for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following
articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  : # Run zombie detector script
./dev-support/zombie-detector.sh --jenkins ${BUILD_ID}
[yahoo-not-h2] $ /bin/bash -xe /tmp/hudson6773435589424810167.sh
+ ./dev-support/zombie-detector.sh --jenkins 1175
Tue Jul  5 23:01:29 UTC 2016 We're ok: there is no zombie test


    {color:green}+1 zombies{color}. No zombie tests found running at the end of the build.
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
[FINDBUGS] Skipping publisher since build result is FAILURE
[CHECKSTYLE] Skipping publisher since build result is FAILURE

Mime
View raw message