From dev-return-105592-archive-asf-public=cust-asf.ponee.io@kafka.apache.org Wed Jul 10 23:40:25 2019 Return-Path: X-Original-To: archive-asf-public@cust-asf.ponee.io Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [207.244.88.153]) by mx-eu-02.ponee.io (Postfix) with SMTP id 7DBA25206EF for ; Thu, 11 Jul 2019 01:40:25 +0200 (CEST) Received: (qmail 21777 invoked by uid 500); 10 Jul 2019 23:39:08 -0000 Mailing-List: contact dev-help@kafka.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@kafka.apache.org Delivered-To: mailing list dev@kafka.apache.org Received: (qmail 21757 invoked by uid 99); 10 Jul 2019 23:39:07 -0000 Received: from Unknown (HELO mailrelay1-lw-us.apache.org) (10.10.3.159) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 10 Jul 2019 23:39:07 +0000 Received: from jenkins02.apache.org (jenkins02.apache.org [195.201.213.130]) by mailrelay1-lw-us.apache.org (ASF Mail Server at mailrelay1-lw-us.apache.org) with ESMTP id 8C23C8214; Wed, 10 Jul 2019 23:39:07 +0000 (UTC) Received: from jenkins02.apache.org (localhost.localdomain [127.0.0.1]) by jenkins02.apache.org (ASF Mail Server at jenkins02.apache.org) with ESMTP id DCE7933E00CC; Wed, 10 Jul 2019 23:39:06 +0000 (UTC) Date: Wed, 10 Jul 2019 23:39:05 +0000 (UTC) From: Apache Jenkins Server To: dev@kafka.apache.org, jason@confluent.io, cmccabe@confluent.io Message-ID: <2089821329.6458.1562801946088.JavaMail.jenkins@jenkins02> In-Reply-To: <1731179569.6361.1562780039857.JavaMail.jenkins@jenkins02> References: <1731179569.6361.1562780039857.JavaMail.jenkins@jenkins02> Subject: Build failed in Jenkins: kafka-trunk-jdk8 #3777 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit X-Instance-Identity: MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAkqVKZPv7YyHBB3FvWfV7XQehwe/Ga3aadzSNknt8g382X3uN8A3SOQ+Ixq9HxS+ZlN6XR4TECySmSRy2JN5Rx8svxAD0TjtSF9LuU98dD+LniNDP7Lq6gvRFuJhbMHoS0nuTizDZLsK4X8TW5MyV9w+jFbdoZfRE5O/Mse0fkOeL5uoIS/3Vvu/W+x9QSjDkB7CaU56bPFlQjqqJBl3Cn9r34CkXQZYnLb/NjW4vcpw0+TgMUAPTIVEr5BTPZRshz19g7huwg3zANT5HBIZnzV4hsVY9w4JHkceFdKi/ibNnjPjsFs9pm0HSGJ/RDxjIvSTYT02eH4+m1RAYaj2E9QIDAQAB X-Jenkins-Job: kafka-trunk-jdk8 X-Jenkins-Result: FAILURE See Changes: [github] MINOR: Only invoke hw update logic for follower fetches (#7064) [github] KAFKA-8634: Update ZooKeeper to 3.5.5 (#6802) ------------------------------------------ [...truncated 6.57 MB...] org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED > Task :connect:basic-auth-extension:test org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testSuccess STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testSuccess PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testBadCredential STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testBadCredential PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testBadPassword STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testBadPassword PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownBearer STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownBearer PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownLoginModule STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownLoginModule PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownCredentialsFile STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testUnknownCredentialsFile PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testEmptyCredentialsFile STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testEmptyCredentialsFile PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testNoFileOption STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testNoFileOption PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testPostWithoutAppropriateCredential STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testPostWithoutAppropriateCredential PASSED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testPostNotChangingConnectorTask STARTED org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > testPostNotChangingConnectorTask PASSED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':connect:runtime:test'. > There were failing tests. See the report at: file:// * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 3h 2m 17s 178 actionable tasks: 175 executed, 3 up-to-date Build step 'Execute shell' marked build as failure [FINDBUGS] Collecting findbugs analysis files... Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1 [FINDBUGS] Searching for all files in that match the pattern **/build/reports/*bugs/*.xml [FINDBUGS] Parsing 17 files in [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. [FINDBUGS] Successfully parsed file with 0 unique warnings and 0 duplicates. Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1 No credentials specified Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1 Using GitBlamer to create author and commit information for all warnings. GIT_COMMIT=d67495d6a7f4c5f7e8736a25d6a11a1c1bef8d87, workspace= [FINDBUGS] Computing warning deltas based on reference build #3774 Recording test results Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1 Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1 Not sending mail to unregistered user noreply@github.com