phoenix-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Apache-Phoenix | origin/4.9-HBase-0.98 | Build Failure
Date Mon, 12 Dec 2016 21:18:47 GMT
<a href="http://phoenix.apache.org"><img src='http://phoenix.apache.org/images/phoenix-logo-small.png'/></a><br><b>origin/4.9-HBase-0.98</b>
branch build status <b>Failure</b>
<hr/>
<b>Source repository</b> https://git-wip-us.apache.org/repos/asf?p=phoenix.git;a=shortlog;h=refs/heads/origin/4.9-HBase-0.98
<hr/>
<b>Compiled Artifacts</b> https://builds.apache.org/job/Phoenix-4.9/lastSuccessfulBuild/artifact/
<hr/>
<b>Test Report</b> https://builds.apache.org/job/Phoenix-4.9/lastCompletedBuild/testReport/
<hr/>
<b>Changes</b><br/>
[samarth] PHOENIX-3518 Memory Leak in RenewLeaseTask

[samarth] TTL on UPGRADE_LOCK cell causes upgrade to fail


<hr/>
<b>Build times for last couple of runs</b><br/>Latest build time is the
right most | Legend blue: normal, red: test failure, gray: timeout<br/>
<img src='https://builds.apache.org/view/M-R/view/Phoenix/job/Phoenix-4.9/buildTimeGraph/png'/>
<br/>
[...truncated 6384 lines...]
Tests run: 26, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 182.777 sec - in org.apache.phoenix.end2end.index.LocalIndexIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.63 sec - in org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT
Running org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.178 sec - in org.apache.phoenix.end2end.index.txn.TxWriteFailureIT
Running org.apache.phoenix.execute.PartialCommitIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.319 sec - in org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.191 sec - in org.apache.phoenix.execute.PartialCommitIT
Running org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT
Running org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.393 sec - in org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT
Running org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT
Running org.apache.phoenix.iterate.RoundRobinResultIteratorWithStatsIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.695 sec - in org.apache.phoenix.iterate.RoundRobinResultIteratorWithStatsIT
Running org.apache.phoenix.iterate.ScannerLeaseRenewalIT
Tests run: 8, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 252.41 sec <<< FAILURE!
- in org.apache.phoenix.end2end.index.MutableIndexFailureIT
testWriteFailureDisablesIndex[MutableIndexFailureIT_transactional=false,localIndex=false,isNamespaceMapped=false](org.apache.phoenix.end2end.index.MutableIndexFailureIT)
 Time elapsed: 8.607 sec  <<< ERROR!
org.apache.phoenix.execute.CommitException: 
org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException:
Failed 2 actions: org.apache.hadoop.hbase.DoNotRetryIOException
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:399)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1013)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1009)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2580)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
: 2 times, servers with issues: priapus.apache.org,54096,1481575994647, priapus.apache.org,35516,1481575994553,

	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:73)
	at org.apache.phoenix.index.PhoenixIndexFailurePolicy.handleFailure(PhoenixIndexFailurePolicy.java:114)
	at org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:151)
	at org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:135)
	at org.apache.phoenix.hbase.index.Indexer.doPostWithExceptions(Indexer.java:453)
	at org.apache.phoenix.hbase.index.Indexer.doPost(Indexer.java:401)
	at org.apache.phoenix.hbase.index.Indexer.postBatchMutate(Indexer.java:396)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$36.call(RegionCoprocessorHost.java:1028)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postBatchMutate(RegionCoprocessorHost.java:1024)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2687)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 2 actions:
org.apache.hadoop.hbase.DoNotRetryIOException
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:399)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1013)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1009)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2580)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
: 2 times, servers with issues: priapus.apache.org,54096,1481575994647, priapus.apache.org,35516,1481575994553,

	at org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:211)
	at org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$500(AsyncProcess.java:195)
	at org.apache.hadoop.hbase.client.AsyncProcess.getErrors(AsyncProcess.java:1082)
	at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processBatchCallback(HConnectionManager.java:2479)
	at org.apache.hadoop.hbase.client.HTable.batchCallback(HTable.java:898)
	at org.apache.hadoop.hbase.client.HTable.batchCallback(HTable.java:913)
	at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:888)
	at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.batch(CoprocessorHost.java:595)
	at org.apache.phoenix.execute.DelegateHTable.batch(DelegateHTable.java:94)
	at org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:169)
	at org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:134)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	... 1 more
: 1 time, servers with issues: priapus.apache.org,50742,1481575994722, 
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT.helpTestWriteFailureDisablesIndex(MutableIndexFailureIT.java:232)
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT.testWriteFailureDisablesIndex(MutableIndexFailureIT.java:134)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: Failed 2 actions: org.apache.hadoop.hbase.DoNotRetryIOException
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:399)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1013)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1009)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2580)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
: 2 times, servers with issues: priapus.apache.org,54096,1481575994647, priapus.apache.org,35516,1481575994553,

	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:73)
	at org.apache.phoenix.index.PhoenixIndexFailurePolicy.handleFailure(PhoenixIndexFailurePolicy.java:114)
	at org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:151)
	at org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:135)
	at org.apache.phoenix.hbase.index.Indexer.doPostWithExceptions(Indexer.java:453)
	at org.apache.phoenix.hbase.index.Indexer.doPost(Indexer.java:401)
	at org.apache.phoenix.hbase.index.Indexer.postBatchMutate(Indexer.java:396)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$36.call(RegionCoprocessorHost.java:1028)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postBatchMutate(RegionCoprocessorHost.java:1024)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2687)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 2 actions:
org.apache.hadoop.hbase.DoNotRetryIOException
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:399)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1013)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1656)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1733)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1688)
	at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1009)
	at org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:2580)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2359)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2314)
	at org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2318)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doBatchOp(HRegionServer.java:4678)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.doNonAtomicRegionMutation(HRegionServer.java:3835)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.multi(HRegionServer.java:3680)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:32500)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2195)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)
: 2 times, servers with issues: priapus.apache.org,54096,1481575994647, priapus.apache.org,35516,1481575994553,

	at org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:211)
	at org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$500(AsyncProcess.java:195)
	at org.apache.hadoop.hbase.client.AsyncProcess.getErrors(AsyncProcess.java:1082)
	at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processBatchCallback(HConnectionManager.java:2479)
	at org.apache.hadoop.hbase.client.HTable.batchCallback(HTable.java:898)
	at org.apache.hadoop.hbase.client.HTable.batchCallback(HTable.java:913)
	at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:888)
	at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.batch(CoprocessorHost.java:595)
	at org.apache.phoenix.execute.DelegateHTable.batch(DelegateHTable.java:94)
	at org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:169)
	at org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:134)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	... 1 more
: 1 time, servers with issues: priapus.apache.org,50742,1481575994722, 
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT.helpTestWriteFailureDisablesIndex(MutableIndexFailureIT.java:232)
	at org.apache.phoenix.end2end.index.MutableIndexFailureIT.testWriteFailureDisablesIndex(MutableIndexFailureIT.java:134)

Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 82.019 sec - in org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT
Running org.apache.phoenix.monitoring.PhoenixMetricsIT
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 79.725 sec - in org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT
Running org.apache.phoenix.rpc.PhoenixClientRpcIT
Running org.apache.phoenix.rpc.PhoenixServerRpcIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.474 sec - in org.apache.phoenix.rpc.PhoenixClientRpcIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.089 sec - in org.apache.phoenix.rpc.PhoenixServerRpcIT
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.688 sec - in org.apache.phoenix.monitoring.PhoenixMetricsIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 169.855 sec - in org.apache.phoenix.iterate.ScannerLeaseRenewalIT

Results :

Tests in error: 
  MutableIndexFailureIT.testWriteFailureDisablesIndex:134->helpTestWriteFailureDisablesIndex:232
ยป Commit

Tests run: 285, Failures: 0, Errors: 1, Skipped: 17

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) @ phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  2.904 s]
[INFO] Phoenix Core ....................................... FAILURE [57:17 min]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix Query Server Client ........................ SKIPPED
[INFO] Phoenix Query Server ............................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 57:22 min
[INFO] Finished at: 2016-12-12T21:00:25+00:00
[INFO] Final Memory: 62M/743M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.19.1:verify
(ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Phoenix-4.9/phoenix-core/target/failsafe-reports
for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following
articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any

Mime
  • Unnamed multipart/mixed (inline, None, 0 bytes)
View raw message