hbase-builds mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: HBase-0.98 #910
Date Thu, 19 Mar 2015 23:03:11 GMT
See <https://builds.apache.org/job/HBase-0.98/910/changes>

Changes:

[matteo.bertozzi] HBASE-13285 Fix flaky getRegions() in TestAccessController.setUp()

------------------------------------------
[...truncated 2114 lines...]
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:693)
	at org.apache.hadoop.util.Shell.runCommand(Shell.java:443)
	at org.apache.hadoop.util.Shell.run(Shell.java:379)
	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:678)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:661)
	at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:639)
	at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:305)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:447)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:357)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:332)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoFileContent(HRegionFileSystem.java:756)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:840)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:803)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.createRegionOnFileSystem(HRegionFileSystem.java:869)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4501)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4471)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4444)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4522)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4402)
	at org.apache.hadoop.hbase.HBaseTestingUtility.createTestRegion(HBaseTestingUtility.java:3439)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testNotCachingDataBlocksDuringCompactionInternals(TestCacheOnWrite.java:429)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction(TestCacheOnWrite.java:485)

testStoreFileCacheOnWrite[106](org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite)  Time elapsed:
0.091 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:693)
	at org.apache.hadoop.util.Shell.runCommand(Shell.java:443)
	at org.apache.hadoop.util.Shell.run(Shell.java:379)
	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:678)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:661)
	at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:639)
	at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:305)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:447)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
	at org.apache.hadoop.fs.FilterFileSystem.create(FilterFileSystem.java:174)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:357)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:332)
	at org.apache.hadoop.hbase.io.hfile.AbstractHFileWriter.createOutputStream(AbstractHFileWriter.java:266)
	at org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:302)
	at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:755)
	at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:706)
	at org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:644)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.writeStoreFile(TestCacheOnWrite.java:384)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testStoreFileCacheOnWriteInternals(TestCacheOnWrite.java:262)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testStoreFileCacheOnWrite(TestCacheOnWrite.java:479)

testNotCachingDataBlocksDuringCompaction[107](org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite)
 Time elapsed: 0.093 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:693)
	at org.apache.hadoop.util.Shell.runCommand(Shell.java:443)
	at org.apache.hadoop.util.Shell.run(Shell.java:379)
	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:678)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:661)
	at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:639)
	at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:305)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:447)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:357)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:332)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoFileContent(HRegionFileSystem.java:756)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:840)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.writeRegionInfoOnFilesystem(HRegionFileSystem.java:803)
	at org.apache.hadoop.hbase.regionserver.HRegionFileSystem.createRegionOnFileSystem(HRegionFileSystem.java:869)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4501)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4471)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4444)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4522)
	at org.apache.hadoop.hbase.regionserver.HRegion.createHRegion(HRegion.java:4402)
	at org.apache.hadoop.hbase.HBaseTestingUtility.createTestRegion(HBaseTestingUtility.java:3439)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testNotCachingDataBlocksDuringCompactionInternals(TestCacheOnWrite.java:429)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction(TestCacheOnWrite.java:485)

testStoreFileCacheOnWrite[107](org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite)  Time elapsed:
0.085 sec  <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:693)
	at org.apache.hadoop.util.Shell.runCommand(Shell.java:443)
	at org.apache.hadoop.util.Shell.run(Shell.java:379)
	at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:678)
	at org.apache.hadoop.util.Shell.execCommand(Shell.java:661)
	at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:639)
	at org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:305)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:447)
	at org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
	at org.apache.hadoop.fs.FilterFileSystem.create(FilterFileSystem.java:174)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:357)
	at org.apache.hadoop.hbase.util.FSUtils.create(FSUtils.java:332)
	at org.apache.hadoop.hbase.io.hfile.AbstractHFileWriter.createOutputStream(AbstractHFileWriter.java:266)
	at org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:302)
	at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:755)
	at org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:706)
	at org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:644)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.writeStoreFile(TestCacheOnWrite.java:384)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testStoreFileCacheOnWriteInternals(TestCacheOnWrite.java:262)
	at org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.testStoreFileCacheOnWrite(TestCacheOnWrite.java:479)

Running org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.422 sec - in org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Running org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.29 sec - in org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 91.123 sec - in org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 194.673 sec - in org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.225 sec - in org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Running org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.88 sec - in org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Running org.apache.hadoop.hbase.filter.TestFilterWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.764 sec - in org.apache.hadoop.hbase.filter.TestFilterWrapper
Running org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.695 sec - in org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.695 sec - in org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter

Results :


Tests in error: 
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:486->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testNotCachingDataBlocksDuringCompaction:485->testNotCachingDataBlocksDuringCompactionInternals:429
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory
  TestCacheOnWrite.testStoreFileCacheOnWrite:479->testStoreFileCacheOnWriteInternals:262->writeStoreFile:384
» OutOfMemory


Tests run: 2227, Failures: 0, Errors: 16, Skipped: 22

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [3.057s]
[INFO] HBase - Checkstyle ................................ SUCCESS [0.654s]
[INFO] HBase - Annotations ............................... SUCCESS [0.848s]
[INFO] HBase - Common .................................... SUCCESS [47.492s]
[INFO] HBase - Protocol .................................. SUCCESS [9.154s]
[INFO] HBase - Client .................................... SUCCESS [50.757s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.945s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [5.995s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [8.108s]
[INFO] HBase - Server .................................... FAILURE [3:29:15.452s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Rest ...................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:31:29.287s
[INFO] Finished at: Thu Mar 19 23:00:34 UTC 2015
[INFO] Final Memory: 51M/683M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18:test (secondPartTestsExecution)
on project hbase-server: There was a timeout or other error in the fork -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following
articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire
but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test"
| grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s):
${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-0.98] $ /bin/bash -xe /tmp/hudson6631607213455244425.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=1
+ [[ 1 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=1
+ [[ 1 != 0 ]]
+ echo 'There are 1 zombie tests, they should have been killed by surefire but survived'
There are 1 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
15292: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Sending artifact delta relative to HBase-0.98 #909
Archived 1703 artifacts
Archive block size is 32768
Received 19 blocks and 283866387 bytes
Compression is 0.2%
Took 1 min 56 sec
Recording test results
Updating HBASE-13285

Mime
View raw message