See 
<https://builds.apache.org/job/kafka-trunk-jdk8/3777/display/redirect?page=changes>

Changes:

[github] MINOR: Only invoke hw update logic for follower fetches (#7064)

[github] KAFKA-8634: Update ZooKeeper to 3.5.5 (#6802)

------------------------------------------
[...truncated 6.57 MB...]
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes
 > testShouldCountClicksPerRegionJava PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes
 > testShouldCountClicksPerRegion STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes
 > testShouldCountClicksPerRegion PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should 
filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should 
filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should 
filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should 
filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join 
correctly records STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join 
correctly records PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a 
Materialized should join correctly records and state store STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a 
Materialized should join correctly records and state store PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress 
should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress 
should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress 
should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress 
should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed 
KTable#suppress should correctly suppress results using 
Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed 
KTable#suppress should correctly suppress results using 
Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed 
KTable#suppress should correctly suppress results using 
Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed 
KTable#suppress should correctly suppress results using 
Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should 
create a Produced with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should 
create a Produced with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with 
timestampExtractor and resetPolicy should create a Consumed with Serdes, 
timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with 
timestampExtractor and resetPolicy should create a Consumed with Serdes, 
timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized 
should create a Materialized with Serdes STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized 
should create a Materialized with Serdes PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a store name should create a Materialized with Serdes and a store name 
STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a store name should create a Materialized with Serdes and a store name 
PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a window store supplier should create a Materialized with Serdes and a 
store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a window store supplier should create a Materialized with Serdes and a 
store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a key value store supplier should create a Materialized with Serdes and a 
store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a key value store supplier should create a Materialized with Serdes and a 
store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a session store supplier should create a Materialized with Serdes and a 
store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize 
with a session store supplier should create a Materialized with Serdes and a 
store supplier PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > 
Suppressed.untilWindowCloses should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > 
Suppressed.untilWindowCloses should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > 
Suppressed.untilTimeLimit should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > 
Suppressed.untilTimeLimit should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords 
should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords 
should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes 
should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes 
should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded 
should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded 
should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should 
support very long chains of factory methods STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should 
support very long chains of factory methods PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should 
create a Consumed with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should 
create a Consumed with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
timestampExtractor and resetPolicy should create a Consumed with Serdes, 
timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
timestampExtractor and resetPolicy should create a Consumed with Serdes, 
timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
timestampExtractor should create a Consumed with Serdes and timestampExtractor 
STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
timestampExtractor should create a Consumed with Serdes and timestampExtractor 
PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
resetPolicy should create a Consumed with Serdes and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with 
resetPolicy should create a Consumed with Serdes and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should 
create a Grouped with Serdes STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should 
create a Grouped with Serdes PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with 
repartition topic name should create a Grouped with Serdes, and repartition 
topic name STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with 
repartition topic name should create a Grouped with Serdes, and repartition 
topic name PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should 
create a Joined with Serdes STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should 
create a Joined with Serdes PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should 
create a Joined with Serdes and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should 
create a Joined with Serdes and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should 
filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should 
filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should 
filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should 
filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should 
run foreach actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should 
run foreach actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run 
peek actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run 
peek actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should 
select a new key STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should 
select a new key PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should 
join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should 
join correctly records PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized 
STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized 
PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED

> Task :connect:basic-auth-extension:test

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testSuccess STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testSuccess PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testBadCredential STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testBadCredential PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testBadPassword STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testBadPassword PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownBearer STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownBearer PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownLoginModule STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownLoginModule PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownCredentialsFile STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testUnknownCredentialsFile PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testEmptyCredentialsFile STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testEmptyCredentialsFile PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testNoFileOption STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testNoFileOption PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testPostWithoutAppropriateCredential STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testPostWithoutAppropriateCredential PASSED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testPostNotChangingConnectorTask STARTED

org.apache.kafka.connect.rest.basic.auth.extension.JaasBasicAuthFilterTest > 
testPostNotChangingConnectorTask PASSED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':connect:runtime:test'.
> There were failing tests. See the report at: 
> file://<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/runtime/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3h 2m 17s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[FINDBUGS] Searching for all files in 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/> that match the pattern 
**/build/reports/*bugs/*.xml
[FINDBUGS] Parsing 17 files in 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/clients/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/api/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/basic-auth-extension/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/file/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/json/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/runtime/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/transforms/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/examples/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/generator/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/jmh-benchmarks/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/log4j-appender/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/examples/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/streams-scala/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/test-utils/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file 
<https://builds.apache.org/job/kafka-trunk-jdk8/ws/tools/build/reports/spotbugs/main.xml>
 with 0 unique warnings and 0 duplicates.
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
No credentials specified
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
<Git Blamer> Using GitBlamer to create author and commit information for all 
warnings.
<Git Blamer> GIT_COMMIT=d67495d6a7f4c5f7e8736a25d6a11a1c1bef8d87, 
workspace=<https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #3774
Recording test results
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Not sending mail to unregistered user nore...@github.com

Reply via email to