kafka-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: Kafka-0.8 #50
Date Tue, 25 Sep 2012 19:24:27 GMT
See <https://builds.apache.org/job/Kafka-0.8/50/changes>

Changes:

[junrao] System Test should remove the top level data log directory; patched by John Fung;
reviewed by Jun Rao; KAFKA-526

------------------------------------------
[...truncated 1253 lines...]
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
	at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
	at java.lang.Runtime.loadLibrary0(Runtime.java:823)
	at java.lang.System.loadLibrary(System.java:1028)
	at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
	... 53 more
[error] Test Failed: testSnappyCompressDecompressExplicit
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
	at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
	at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
	at kafka.message.SnappyCompression.<init>(CompressionUtils.scala:61)
	at kafka.message.CompressionFactory$.apply(CompressionUtils.scala:82)
	at kafka.message.CompressionUtils$.compress(CompressionUtils.scala:109)
	at kafka.message.CompressionUtilTest.testSnappyCompressDecompressExplicit(CompressionUtilsTest.scala:65)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] Test Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] Test Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] Test Starting: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Starting: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Passed: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] Test Starting: testSerializationAndDeserialization
[info] Test Passed: testSerializationAndDeserialization
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest
==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest
==
[info] 
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] Test Starting: testAppend
[info] Test Passed: testAppend
[info] Test Starting: testTrunc
[info] Test Passed: testTrunc
[info] Test Starting: testTruncLast
[info] Test Passed: testTruncLast
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[info] Test Passed: testTimeBasedLogRoll
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testLoadInvalidLogsFails
[info] Test Passed: testLoadInvalidLogsFails
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRolls
[info] Test Passed: testEdgeLogRolls
[info] Test Starting: testMessageSizeCheck
[info] Test Passed: testMessageSizeCheck
[info] Test Starting: testTruncateTo
[info] Test Passed: testTruncateTo
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] Test Starting: testWhitelists
[info] Test Passed: testWhitelists
[info] Test Starting: testBlacklists
[info] Test Passed: testBlacklists
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 136, Failed 1, Errors 0, Passed 135, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /var/tmp/sbt_37b84a09
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == perf / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == perf / test-compile ==
[info] 
[info] == hadoop consumer / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == hadoop consumer / test-compile ==
[info] 
[info] == hadoop producer / copy-test-resources ==
[info] == hadoop producer / copy-test-resources ==
[info] 
[info] == hadoop producer / copy-resources ==
[info] == hadoop producer / copy-resources ==
[info] 
[info] == java-examples / copy-resources ==
[info] == java-examples / copy-resources ==
[info] 
[info] == perf / copy-test-resources ==
[info] == perf / copy-test-resources ==
[error] Error running kafka.message.CompressionUtilTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 204 s, completed Sep 25, 2012 7:24:48 PM
[info] 
[info] Total session time: 205 s, completed Sep 25, 2012 7:24:48 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Mime
View raw message