[jira] [Assigned] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Egerton reassigned KAFKA-9258:


Assignee: (was: Chris Egerton)

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Egerton updated KAFKA-9258:
-
Affects Version/s: 2.4.0

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Affects Versions: 2.4.0
>Reporter: Cyrus Vafadari
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Egerton reassigned KAFKA-9258:


Assignee: (was: Chris Egerton)

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Egerton reassigned KAFKA-9258:


Assignee: Chris Egerton

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Assignee: Chris Egerton
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986557#comment-16986557
 ] 

Chris Egerton commented on KAFKA-9258:
--

This issue prevents users from restarting tasks that have failed during 
startup, and causes the REST API to throw a 500 error if they try to do so. 
This persists even after the connector has been reconfigured.

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Chris Egerton (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Egerton reassigned KAFKA-9258:


Assignee: Chris Egerton

> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Assignee: Chris Egerton
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9259) suppress() for windowed-Serdes does not work with default serdes

2019-12-02 Thread John Roesler (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986546#comment-16986546
 ] 

John Roesler commented on KAFKA-9259:
-

Thanks for the report, [~mjsax]!

Just a quick additional note for whoever picks this up: To properly 
parameterize windowed serdes, you need extra information about the windows (eg 
the window size). Theoretically, the suppress operator could find this out by 
traversing the topology upstream to find it on its window-definition ancestor, 
but it might get a bit brittle. If this turns out to be a mess, it would 
probably be cleaner just to implement 
https://issues.apache.org/jira/browse/KAFKA-9260 instead, which would also fix 
this issue.

> suppress() for windowed-Serdes does not work with default serdes
> 
>
> Key: KAFKA-9259
> URL: https://issues.apache.org/jira/browse/KAFKA-9259
> Project: Kafka
>  Issue Type: Bug
>  Components: streams
>Affects Versions: 2.1.0
>Reporter: Matthias J. Sax
>Priority: Major
>
> The suppress() operator either inherits serdes from its upstream operator or 
> falls back to default serdes from the config.
> If the upstream operator is an windowed aggregation, the window-aggregation 
> operator wraps the user passed-in serde with a window-serde and pushed it 
> into suppress() – however, if default serdes are used, the window-aggregation 
> operator cannot push anything into suppress(). At runtime, it just creates a 
> default serde and wraps it according. For this case, suppress() also falls 
> back to default serdes; however, it does not wrap the serde and thus a 
> ClassCastException is thrown when the serde is used later.
> suppress() is already aware if the upstream aggregation is time/session 
> windowed or not and thus should use this information to wrap default serdes 
> accordingly.
> The current workaround for windowed-suppress is to overwrite the default 
> serde upstream to suppress(), such that suppress() inherits serdes and does 
> not fall back to default serdes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (KAFKA-9254) Topic level configuration failed

2019-12-02 Thread fenghong (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986025#comment-16986025
 ] 

fenghong edited comment on KAFKA-9254 at 12/3/19 1:48 AM:
--

We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Write data continuously during testing

 

producer config
{code:java}
acks=all
{code}
 OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
server.properties
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

 

 


was (Author: fenghong):
We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Write data continuously during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
server.properties
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

 

 

> Topic level configuration failed
> 
>
> Key: KAFKA-9254
> URL: https://issues.apache.org/jira/browse/KAFKA-9254
> Project: Kafka
>  Issue Type: Bug
>  Components: config, log, replication
>Affects Versions: 2.0.1
>Reporter: fenghong
>Priority: Critical
>
> We are engineers at Huobi and now encounter Kafka BUG 
> Modifying DynamicBrokerConfig more than 2 times will invalidate the topic 
> level unrelated configuration
> The bug reproduction method as follows:
>  # Set Kafka Broker config  server.properties min.insync.replicas=3
>  # Create topic test-1 and set topic‘s level config min.insync.replicas=2
>  # Dynamically modify the configuration twice as shown below
> {code:java}
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.message.timestamp.type=LogAppendTime
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.retention.ms=60480
> {code}
>  # stop a Kafka Server and found the Exception as shown below
>  org.apache.kafka.common.errors.NotEnoughReplicasException: Number of insync 
> replicas for partition test-1-0 is [2], below required minimum [3]
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KAFKA-9261) NPE when updating client metadata

2019-12-02 Thread Jason Gustafson (Jira)
Jason Gustafson created KAFKA-9261:
--

 Summary: NPE when updating client metadata
 Key: KAFKA-9261
 URL: https://issues.apache.org/jira/browse/KAFKA-9261
 Project: Kafka
  Issue Type: Bug
Reporter: Jason Gustafson
Assignee: Jason Gustafson


We have seen the following exception recently:

{code}
java.lang.NullPointerException
at java.base/java.util.Objects.requireNonNull(Objects.java:221)
at org.apache.kafka.common.Cluster.(Cluster.java:134)
at org.apache.kafka.common.Cluster.(Cluster.java:89)
at 
org.apache.kafka.clients.MetadataCache.computeClusterView(MetadataCache.java:120)
at org.apache.kafka.clients.MetadataCache.(MetadataCache.java:82)
at org.apache.kafka.clients.MetadataCache.(MetadataCache.java:58)
at 
org.apache.kafka.clients.Metadata.handleMetadataResponse(Metadata.java:325)
at org.apache.kafka.clients.Metadata.update(Metadata.java:252)
at 
org.apache.kafka.clients.NetworkClient$DefaultMetadataUpdater.handleCompletedMetadataResponse(NetworkClient.java:1059)
at 
org.apache.kafka.clients.NetworkClient.handleCompletedReceives(NetworkClient.java:845)
at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:548)
at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:262)
at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:233)
at 
org.apache.kafka.clients.consumer.KafkaConsumer.pollForFetches(KafkaConsumer.java:1281)
at 
org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1225)
at 
org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:1201)
{code}

The client assumes that if a leader is included in the response, then node 
information must also be available. There are at least a couple possible 
reasons this assumption can fail:

1. The client is able to detect stale partition metadata using leader epoch 
information available. If stale partition metadata is detected, the client 
ignores it and uses the last known metadata. However, it cannot detect stale 
broker information and will always accept the latest update. This means that 
the latest metadata may be a mix of multiple metadata responses and therefore 
the invariant will not generally hold.
2. There is no lock which protects both the fetching of partition metadata and 
the live broker when handling a Metadata request. This means an UpdateMetadata 
request can arrive concurrently and break the intended invariant.

It seems case 2 has been possible for a long time, but it should be extremely 
rare. Case 1 was only made possible with KIP-320, which added the leader epoch 
tracking. It should also be rare, but the window for inconsistent metadata is 
probably a bit bigger than the window for a concurrent update.

To fix this, we should make the client more defensive about metadata updates 
and not assume that the leader is among the live endpoints.





--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9210) kafka stream loss data

2019-12-02 Thread Sophie Blee-Goldman (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986466#comment-16986466
 ] 

Sophie Blee-Goldman commented on KAFKA-9210:


Hi [~panpan.liu], it seems you might be hitting KAFKA-7443. That issue has been 
fixed in 2.1.1 so I suggest upgrading your Streams client, and verifying that 
you are getting the expected results then

> kafka stream loss data
> --
>
> Key: KAFKA-9210
> URL: https://issues.apache.org/jira/browse/KAFKA-9210
> Project: Kafka
>  Issue Type: Bug
>  Components: streams
>Affects Versions: 2.0.1
>Reporter: panpan.liu
>Priority: Major
> Attachments: app.log, screenshot-1.png
>
>
> kafka broker: 2.0.1
> kafka stream client: 2.1.0
>  # two applications run at the same time
>  # after some days,I stop one application(in k8s)
>  # The flollowing log occured and I check the data and find that value is 
> less than what I expected.
> {quote}Partitions 
> [flash-app-xmc-worker-share-store-minute-repartition-1]Partitions 
> [flash-app-xmc-worker-share-store-minute-repartition-1] for changelog 
> flash-app-xmc-worker-share-store-minute-changelog-12019-11-19 
> 05:50:49.816|WARN 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|101|stream-thread
>  [flash-client-xmc-StreamThread-3] Restoring StreamTasks failed. Deleting 
> StreamTasks stores to recreate from 
> scratch.org.apache.kafka.clients.consumer.OffsetOutOfRangeException: Offsets 
> out of range with no configured reset policy for partitions: 
> \{flash-app-xmc-worker-share-store-minute-changelog-1=6128684} at 
> org.apache.kafka.clients.consumer.internals.Fetcher.parseCompletedFetch(Fetcher.java:987)2019-11-19
>  05:50:49.817|INFO 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|105|stream-thread
>  [flash-client-xmc-StreamThread-3] Reinitializing StreamTask TaskId: 10_1 
> ProcessorTopology: KSTREAM-SOURCE-70: topics: 
> [flash-app-xmc-worker-share-store-minute-repartition] children: 
> [KSTREAM-AGGREGATE-67] KSTREAM-AGGREGATE-67: states: 
> [worker-share-store-minute] children: [KTABLE-TOSTREAM-71] 
> KTABLE-TOSTREAM-71: children: [KSTREAM-SINK-72] 
> KSTREAM-SINK-72: topic: 
> StaticTopicNameExtractor(xmc-worker-share-minute)Partitions 
> [flash-app-xmc-worker-share-store-minute-repartition-1] for changelog 
> flash-app-xmc-worker-share-store-minute-changelog-12019-11-19 
> 05:50:49.842|WARN 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|101|stream-thread
>  [flash-client-xmc-StreamThread-3] Restoring StreamTasks failed. Deleting 
> StreamTasks stores to recreate from 
> scratch.org.apache.kafka.clients.consumer.OffsetOutOfRangeException: Offsets 
> out of range with no configured reset policy for partitions: 
> \{flash-app-xmc-worker-share-store-minute-changelog-1=6128684} at 
> org.apache.kafka.clients.consumer.internals.Fetcher.parseCompletedFetch(Fetcher.java:987)2019-11-19
>  05:50:49.842|INFO 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|105|stream-thread
>  [flash-client-xmc-StreamThread-3] Reinitializing StreamTask TaskId: 10_1 
> ProcessorTopology: KSTREAM-SOURCE-70: topics: 
> [flash-app-xmc-worker-share-store-minute-repartition] children: 
> [KSTREAM-AGGREGATE-67] KSTREAM-AGGREGATE-67: states: 
> [worker-share-store-minute] children: [KTABLE-TOSTREAM-71] 
> KTABLE-TOSTREAM-71: children: [KSTREAM-SINK-72] 
> KSTREAM-SINK-72: topic: 
> StaticTopicNameExtractor(xmc-worker-share-minute)Partitions 
> [flash-app-xmc-worker-share-store-minute-repartition-1] for changelog 
> flash-app-xmc-worker-share-store-minute-changelog-12019-11-19 
> 05:50:49.905|WARN 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|101|stream-thread
>  [flash-client-xmc-StreamThread-3] Restoring StreamTasks failed. Deleting 
> StreamTasks stores to recreate from 
> scratch.org.apache.kafka.clients.consumer.OffsetOutOfRangeException: Offsets 
> out of range with no configured reset policy for partitions: 
> \{flash-app-xmc-worker-share-store-minute-changelog-1=6128684} at 
> org.apache.kafka.clients.consumer.internals.Fetcher.parseCompletedFetch(Fetcher.java:987)2019-11-19
>  05:50:49.906|INFO 
> |flash-client-xmc-StreamThread-3|o.a.k.s.p.i.StoreChangelogReader|105|stream-thread
>  [flash-client-xmc-StreamThread-3] Reinitializing StreamTask TaskId: 10_1 
> ProcessorTopology: KSTREAM-SOURCE-70: topics: 
> [flash-app-xmc-worker-share-store-minute-repartition] children: 
> [KSTREAM-AGGREGATE-67] KSTREAM-AGGREGATE-67: states: 
> [worker-share-store-minute] 
> {quote}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9173) StreamsPartitionAssignor assigns partitions to only one worker

2019-12-02 Thread Sophie Blee-Goldman (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9173?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986460#comment-16986460
 ] 

Sophie Blee-Goldman commented on KAFKA-9173:


I agree we should aim to fix this, but the workaround here is pretty simple – 
you just need to decrease the number of threads to better match the actual 
workload. If your app only has 10 tasks and you want to run 10 instances, each 
instance only needs one thread. Even if we did "fix" the assignor so that it 
spread the 10 tasks evenly across the 10 instances, 19 of the 20 threads would 
have no tasks assigned and nothing to do.

You can change num.threads to 1 for each instance, and you will see each be 
assigned one of the ten tasks

> StreamsPartitionAssignor assigns partitions to only one worker
> --
>
> Key: KAFKA-9173
> URL: https://issues.apache.org/jira/browse/KAFKA-9173
> Project: Kafka
>  Issue Type: Bug
>  Components: streams
>Affects Versions: 2.3.0, 2.2.1
>Reporter: Oleg Muravskiy
>Priority: Major
>  Labels: user-experience
> Attachments: StreamsPartitionAssignor.log
>
>
> I'm running a distributed KafkaStreams application on 10 worker nodes, 
> subscribed to 21 topics with 10 partitions in each. I'm only using a 
> Processor interface, and a persistent state store.
> However, only one worker gets assigned partitions, all other workers get 
> nothing. Restarting the application, or cleaning local state stores does not 
> help. StreamsPartitionAssignor migrates to other nodes, and eventually picks 
> up other node to assign partitions to, but still only one node.
> It's difficult to figure out where to look for the signs of problems, I'm 
> attaching the log messages from the StreamsPartitionAssignor. Let me know 
> what else I could provide to help resolve this.
> [^StreamsPartitionAssignor.log]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KAFKA-9225) kafka fail to run on linux-aarch64

2019-12-02 Thread Sophie Blee-Goldman (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sophie Blee-Goldman updated KAFKA-9225:
---
Fix Version/s: 3.0.0

> kafka fail to run on linux-aarch64
> --
>
> Key: KAFKA-9225
> URL: https://issues.apache.org/jira/browse/KAFKA-9225
> Project: Kafka
>  Issue Type: Bug
>  Components: streams
>Reporter: jiamei xie
>Priority: Major
>  Labels: incompatible
> Fix For: 3.0.0
>
> Attachments: compat_report.html
>
>
> *Steps to reproduce:*
> 1. Download Kafka latest source code
> 2. Build it with gradle
> 3. Run 
> [streamDemo|[https://kafka.apache.org/23/documentation/streams/quickstart]]
> when running bin/kafka-run-class.sh 
> org.apache.kafka.streams.examples.wordcount.WordCountDemo, it crashed with 
> the following error message
> {code:java}
> xjm@ubuntu-arm01:~/kafka$bin/kafka-run-class.sh 
> org.apache.kafka.streams.examples.wordcount.WordCountDemo
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/core/build/dependant-libs-2.12.10/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/tools/build/dependant-libs-2.12.10/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/api/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/transforms/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/runtime/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/file/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/mirror/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/mirror-client/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/json/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/basic-auth-extension/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> [2019-11-19 15:42:23,277] WARN The configuration 'admin.retries' was supplied 
> but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
> [2019-11-19 15:42:23,278] WARN The configuration 'admin.retry.backoff.ms' was 
> supplied but isn't a known config. 
> (org.apache.kafka.clients.consumer.ConsumerConfig)
> [2019-11-19 15:42:24,278] ERROR stream-client 
> [streams-wordcount-0f3cf88b-e2c4-4fb6-b7a3-9754fad5cd48] All stream threads 
> have died. The instance will be in error state and should be closed. 
> (org.apach e.kafka.streams.KafkaStreams)
> Exception in thread 
> "streams-wordcount-0f3cf88b-e2c4-4fb6-b7a3-9754fad5cd48-StreamThread-1" 
> java.lang.UnsatisfiedLinkError: /tmp/librocksdbjni1377754636857652484.so: 
> /tmp/librocksdbjni13777546368576524 84.so: 
> cannot open shared object file: No such file or directory (Possible cause: 
> can't load AMD 64-bit .so on a AARCH64-bit platform)
> {code}
>  *Analyze:*
> This issue is caused by rocksdbjni-5.18.3.jar which doesn't come with aarch64 
> native support. Replace rocksdbjni-5.18.3.jar with rocksdbjni-6.3.6.jar from 
> [https://mvnrepository.com/artifact/org.rocksdb/rocksdbjni/6.3.6] can fix 
> this problem.
> Attached is the binary compatibility report of rocksdbjni.jar between 5.18.3 
> and 6.3.6. The result is 81.8%. So is it possible to upgrade rocksdbjni to 
> 6.3.6 in upstream? Should there be any kind of tests to execute, please 
> kindly point me. Thanks a lot.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (KAFKA-9225) kafka fail to run on linux-aarch64

2019-12-02 Thread Sophie Blee-Goldman (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sophie Blee-Goldman updated KAFKA-9225:
---
Priority: Blocker  (was: Major)

> kafka fail to run on linux-aarch64
> --
>
> Key: KAFKA-9225
> URL: https://issues.apache.org/jira/browse/KAFKA-9225
> Project: Kafka
>  Issue Type: Bug
>  Components: streams
>Reporter: jiamei xie
>Priority: Blocker
>  Labels: incompatible
> Fix For: 3.0.0
>
> Attachments: compat_report.html
>
>
> *Steps to reproduce:*
> 1. Download Kafka latest source code
> 2. Build it with gradle
> 3. Run 
> [streamDemo|[https://kafka.apache.org/23/documentation/streams/quickstart]]
> when running bin/kafka-run-class.sh 
> org.apache.kafka.streams.examples.wordcount.WordCountDemo, it crashed with 
> the following error message
> {code:java}
> xjm@ubuntu-arm01:~/kafka$bin/kafka-run-class.sh 
> org.apache.kafka.streams.examples.wordcount.WordCountDemo
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/core/build/dependant-libs-2.12.10/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/tools/build/dependant-libs-2.12.10/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/api/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/transforms/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/runtime/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/file/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/mirror/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/mirror-client/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/json/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/home/xjm/kafka/connect/basic-auth-extension/build/dependant-libs/slf4j-log4j12-1.7.28.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> [2019-11-19 15:42:23,277] WARN The configuration 'admin.retries' was supplied 
> but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
> [2019-11-19 15:42:23,278] WARN The configuration 'admin.retry.backoff.ms' was 
> supplied but isn't a known config. 
> (org.apache.kafka.clients.consumer.ConsumerConfig)
> [2019-11-19 15:42:24,278] ERROR stream-client 
> [streams-wordcount-0f3cf88b-e2c4-4fb6-b7a3-9754fad5cd48] All stream threads 
> have died. The instance will be in error state and should be closed. 
> (org.apach e.kafka.streams.KafkaStreams)
> Exception in thread 
> "streams-wordcount-0f3cf88b-e2c4-4fb6-b7a3-9754fad5cd48-StreamThread-1" 
> java.lang.UnsatisfiedLinkError: /tmp/librocksdbjni1377754636857652484.so: 
> /tmp/librocksdbjni13777546368576524 84.so: 
> cannot open shared object file: No such file or directory (Possible cause: 
> can't load AMD 64-bit .so on a AARCH64-bit platform)
> {code}
>  *Analyze:*
> This issue is caused by rocksdbjni-5.18.3.jar which doesn't come with aarch64 
> native support. Replace rocksdbjni-5.18.3.jar with rocksdbjni-6.3.6.jar from 
> [https://mvnrepository.com/artifact/org.rocksdb/rocksdbjni/6.3.6] can fix 
> this problem.
> Attached is the binary compatibility report of rocksdbjni.jar between 5.18.3 
> and 6.3.6. The result is 81.8%. So is it possible to upgrade rocksdbjni to 
> 6.3.6 in upstream? Should there be any kind of tests to execute, please 
> kindly point me. Thanks a lot.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KAFKA-9260) Improve Serde "push down" and "wrapping"

2019-12-02 Thread Matthias J. Sax (Jira)
Matthias J. Sax created KAFKA-9260:
--

 Summary: Improve Serde "push down" and "wrapping"
 Key: KAFKA-9260
 URL: https://issues.apache.org/jira/browse/KAFKA-9260
 Project: Kafka
  Issue Type: Improvement
  Components: streams
Reporter: Matthias J. Sax


Kafka Streams DSL supports "serde push down" feature to let downstream 
operators inherit upstream serdes if key and/or value are not modified. 
Furthermore, some operators "wrap" user specified serdes internally (eg, 
windowed aggregations wrap the user specified key-serde with a time/session 
window serde – some stores use `ValueAndTimestampSerde` and foreign-key joins 
also uses some internal wrappers).

The current implementation faces couple of issues, because the "serde push 
down" feature is a DSL level feature that is used when the Topology is 
generated. Furthermore, "serde wrapping" is an operator specific feature, not a 
DSL concept per-se. At runtime, neither "push down" nor "wrapping" are know 
concepts.

This design implies that if users specify serdes, wrapping and push down works 
as expected. However, if we fall back to default serdes, special care needs to 
be taken: for example, some operators not apply the wrapping logic during 
translation time, and there is additional code that does the wrapping of 
default serdes as runtime. Another approach would be to wrap a null-Serde, and 
update the wrapper later (ie, overwrite `null` with the default serde from the 
config).

Overall, the current design leads to bugs (eg, KAFKA-9248 and KAFKA-9259), and 
user confusion how it actually works and when/where to specify serdes. Hence, 
we should consider to rework how we do serde push down and/or wrapping.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KAFKA-9259) suppress() for windowed-Serdes does not work with default serdes

2019-12-02 Thread Matthias J. Sax (Jira)
Matthias J. Sax created KAFKA-9259:
--

 Summary: suppress() for windowed-Serdes does not work with default 
serdes
 Key: KAFKA-9259
 URL: https://issues.apache.org/jira/browse/KAFKA-9259
 Project: Kafka
  Issue Type: Bug
  Components: streams
Affects Versions: 2.1.0
Reporter: Matthias J. Sax


The suppress() operator either inherits serdes from its upstream operator or 
falls back to default serdes from the config.

If the upstream operator is an windowed aggregation, the window-aggregation 
operator wraps the user passed-in serde with a window-serde and pushed it into 
suppress() – however, if default serdes are used, the window-aggregation 
operator cannot push anything into suppress(). At runtime, it just creates a 
default serde and wraps it according. For this case, suppress() also falls back 
to default serdes; however, it does not wrap the serde and thus a 
ClassCastException is thrown when the serde is used later.

suppress() is already aware if the upstream aggregation is time/session 
windowed or not and thus should use this information to wrap default serdes 
accordingly.

The current workaround for windowed-suppress is to overwrite the default serde 
upstream to suppress(), such that suppress() inherits serdes and does not fall 
back to default serdes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9071) transactional.id.expiration.ms config value should be implemented as a Long

2019-12-02 Thread Mario Georgiev (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986434#comment-16986434
 ] 

Mario Georgiev commented on KAFKA-9071:
---

Can we move the ticket to review?

> transactional.id.expiration.ms config value should be implemented as a Long
> ---
>
> Key: KAFKA-9071
> URL: https://issues.apache.org/jira/browse/KAFKA-9071
> Project: Kafka
>  Issue Type: Improvement
>  Components: config
>Affects Versions: 2.3.0
>Reporter: Joost van de Wijgerd
>Assignee: Mario Georgiev
>Priority: Major
>
> Currently the value of this config parameter is limited to MAX_INT 
> effectively limiting the transactional id expiration to  ~ 25 days. This is 
> causing some issues for us on our Acceptance environment (which is not used 
> that often / heavily) where our transactional services will start failing 
> because if this issue.
> I believe best practice for millisecond values should be to implement them as 
> a Long and not as an Integer
> this is currently the max value: transactional.id.expiration.ms=2147483647
> while I would like to set it to: transactional.id.expiration.ms=3154000 
> (i.e. 1 year)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986386#comment-16986386
 ] 

ASF GitHub Bot commented on KAFKA-9258:
---

cyrusv commented on pull request #7768: KAFKA-9258 Check Connect Metrics 
non-null in task stop
URL: https://github.com/apache/kafka/pull/7768
 
 
   Connect sometimes will stop task when start() has failed. In these cases, we 
must guard against NPE as noted in 
https://issues.apache.org/jira/browse/KAFKA-9258
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Connect ConnectorStatusMetricsGroup sometimes NPE
> -
>
> Key: KAFKA-9258
> URL: https://issues.apache.org/jira/browse/KAFKA-9258
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Cyrus Vafadari
>Priority: Major
>
> java.lang.NullPointerException
>   at 
> org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
>   at 
> org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
>   at 
> org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
>   at 
> org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
>   at java.util.concurrent.FutureTask.run(FutureTask.java)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (KAFKA-9258) Connect ConnectorStatusMetricsGroup sometimes NPE

2019-12-02 Thread Cyrus Vafadari (Jira)
Cyrus Vafadari created KAFKA-9258:
-

 Summary: Connect ConnectorStatusMetricsGroup sometimes NPE
 Key: KAFKA-9258
 URL: https://issues.apache.org/jira/browse/KAFKA-9258
 Project: Kafka
  Issue Type: Bug
  Components: KafkaConnect
Reporter: Cyrus Vafadari


java.lang.NullPointerException
at 
org.apache.kafka.connect.runtime.Worker$ConnectorStatusMetricsGroup.recordTaskRemoved(Worker.java:901)
at 
org.apache.kafka.connect.runtime.Worker.awaitStopTask(Worker.java:720)
at 
org.apache.kafka.connect.runtime.Worker.awaitStopTasks(Worker.java:740)
at 
org.apache.kafka.connect.runtime.Worker.stopAndAwaitTasks(Worker.java:758)
at 
org.apache.kafka.connect.runtime.distributed.DistributedHerder.processTaskConfigUpdatesWithIncrementalCooperative(DistributedHerder.java:575)
at 
org.apache.kafka.connect.runtime.distributed.DistributedHerder.tick(DistributedHerder.java:396)
at 
org.apache.kafka.connect.runtime.distributed.DistributedHerder.run(DistributedHerder.java:282)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run$$$capture(FutureTask.java:266)
at java.util.concurrent.FutureTask.run(FutureTask.java)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9071) transactional.id.expiration.ms config value should be implemented as a Long

2019-12-02 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986351#comment-16986351
 ] 

ASF GitHub Bot commented on KAFKA-9071:
---

despondency commented on pull request #7767: KAFKA-9071 
transactional.id.expiration.ms config value should be impl…
URL: https://github.com/apache/kafka/pull/7767
 
 
   transactional.id.expiration.ms config value should be implemented as a Long
   
   ### Committer Checklist (excluded from commit message)
   - [ ] Verify design and implementation 
   - [ ] Verify test coverage and CI build status
   - [ ] Verify documentation (including upgrade notes)
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> transactional.id.expiration.ms config value should be implemented as a Long
> ---
>
> Key: KAFKA-9071
> URL: https://issues.apache.org/jira/browse/KAFKA-9071
> Project: Kafka
>  Issue Type: Improvement
>  Components: config
>Affects Versions: 2.3.0
>Reporter: Joost van de Wijgerd
>Assignee: Mario Georgiev
>Priority: Major
>
> Currently the value of this config parameter is limited to MAX_INT 
> effectively limiting the transactional id expiration to  ~ 25 days. This is 
> causing some issues for us on our Acceptance environment (which is not used 
> that often / heavily) where our transactional services will start failing 
> because if this issue.
> I believe best practice for millisecond values should be to implement them as 
> a Long and not as an Integer
> this is currently the max value: transactional.id.expiration.ms=2147483647
> while I would like to set it to: transactional.id.expiration.ms=3154000 
> (i.e. 1 year)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9071) transactional.id.expiration.ms config value should be implemented as a Long

2019-12-02 Thread Mario Georgiev (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mario Georgiev reassigned KAFKA-9071:
-

Assignee: Mario Georgiev

> transactional.id.expiration.ms config value should be implemented as a Long
> ---
>
> Key: KAFKA-9071
> URL: https://issues.apache.org/jira/browse/KAFKA-9071
> Project: Kafka
>  Issue Type: Improvement
>  Components: config
>Affects Versions: 2.3.0
>Reporter: Joost van de Wijgerd
>Assignee: Mario Georgiev
>Priority: Major
>
> Currently the value of this config parameter is limited to MAX_INT 
> effectively limiting the transactional id expiration to  ~ 25 days. This is 
> causing some issues for us on our Acceptance environment (which is not used 
> that often / heavily) where our transactional services will start failing 
> because if this issue.
> I believe best practice for millisecond values should be to implement them as 
> a Long and not as an Integer
> this is currently the max value: transactional.id.expiration.ms=2147483647
> while I would like to set it to: transactional.id.expiration.ms=3154000 
> (i.e. 1 year)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9156) LazyTimeIndex & LazyOffsetIndex may cause niobufferoverflow in concurrent state

2019-12-02 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986321#comment-16986321
 ] 

ASF GitHub Bot commented on KAFKA-9156:
---

ijuma commented on pull request #7760: KAFKA-9156: fix LazyTimeIndex & 
LazyOffsetIndex concurrency
URL: https://github.com/apache/kafka/pull/7760
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> LazyTimeIndex & LazyOffsetIndex may cause niobufferoverflow in concurrent 
> state
> ---
>
> Key: KAFKA-9156
> URL: https://issues.apache.org/jira/browse/KAFKA-9156
> Project: Kafka
>  Issue Type: Bug
>  Components: core
>Affects Versions: 2.3.0, 2.3.1
>Reporter: shilin Lu
>Assignee: Alex Mironov
>Priority: Blocker
>  Labels: regression
> Fix For: 2.4.0, 2.3.2
>
> Attachments: image-2019-11-07-17-42-13-852.png, 
> image-2019-11-07-17-44-05-357.png, image-2019-11-07-17-46-53-650.png
>
>
> !image-2019-11-07-17-42-13-852.png!
> this timeindex get function is not thread safe ,may cause create some 
> timeindex.
> !image-2019-11-07-17-44-05-357.png!
> When create timeindex not exactly one ,may cause mappedbytebuffer position to 
> end. Then write index entry to this mmap file will cause 
> java.nio.BufferOverflowException.
>  
> !image-2019-11-07-17-46-53-650.png!
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-5034) Connect workers don't discover new Connector Plugins without Restart

2019-12-02 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-5034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986252#comment-16986252
 ] 

ASF GitHub Bot commented on KAFKA-5034:
---

C0urante commented on pull request #4905: KAFKA-5034: Enable plugins to be 
added to the plugin path at runtime
URL: https://github.com/apache/kafka/pull/4905
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Connect workers don't discover new Connector Plugins without Restart
> 
>
> Key: KAFKA-5034
> URL: https://issues.apache.org/jira/browse/KAFKA-5034
> Project: Kafka
>  Issue Type: Bug
>  Components: KafkaConnect
>Reporter: Gwen Shapira
>Priority: Major
>
> If I want to add a new Connector Plugin to a running distributed Connect 
> cluster, I need to copy the JAR to the classpath and then restart all the 
> workers so they will pick up the new plugin before I can create a connector.
> This is both un-intuitive (most modern up can pick up changes dynamically) 
> and can get difficult when a connect cluster is shared between teams.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9156) LazyTimeIndex & LazyOffsetIndex may cause niobufferoverflow in concurrent state

2019-12-02 Thread Alex Mironov (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alex Mironov reassigned KAFKA-9156:
---

Assignee: Alex Mironov

> LazyTimeIndex & LazyOffsetIndex may cause niobufferoverflow in concurrent 
> state
> ---
>
> Key: KAFKA-9156
> URL: https://issues.apache.org/jira/browse/KAFKA-9156
> Project: Kafka
>  Issue Type: Bug
>  Components: core
>Affects Versions: 2.3.0, 2.3.1
>Reporter: shilin Lu
>Assignee: Alex Mironov
>Priority: Blocker
>  Labels: regression
> Fix For: 2.4.0, 2.3.2
>
> Attachments: image-2019-11-07-17-42-13-852.png, 
> image-2019-11-07-17-44-05-357.png, image-2019-11-07-17-46-53-650.png
>
>
> !image-2019-11-07-17-42-13-852.png!
> this timeindex get function is not thread safe ,may cause create some 
> timeindex.
> !image-2019-11-07-17-44-05-357.png!
> When create timeindex not exactly one ,may cause mappedbytebuffer position to 
> end. Then write index entry to this mmap file will cause 
> java.nio.BufferOverflowException.
>  
> !image-2019-11-07-17-46-53-650.png!
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (KAFKA-9164) Don't evict active topics' metadata from the producer's cache

2019-12-02 Thread Brian Byrne (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brian Byrne resolved KAFKA-9164.

Resolution: Invalid

Marking issue invalid since the producer logic was actually handling this 
correctly.

> Don't evict active topics' metadata from the producer's cache
> -
>
> Key: KAFKA-9164
> URL: https://issues.apache.org/jira/browse/KAFKA-9164
> Project: Kafka
>  Issue Type: Bug
>  Components: producer 
>Reporter: Brian Byrne
>Assignee: Brian Byrne
>Priority: Minor
>
> The producer's metadata currently marks a topic as "not needing to be 
> retained" if it has been 5 minutes since it was first considered, regardless 
> of whether records were being actively produced for the topic. This shouldn't 
> happen and should be fixed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (KAFKA-9255) MessageSet v1 protocol wrong specification

2019-12-02 Thread Mickael Maison (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9255?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mickael Maison reassigned KAFKA-9255:
-

Assignee: Fábio Silva

> MessageSet v1 protocol wrong specification
> --
>
> Key: KAFKA-9255
> URL: https://issues.apache.org/jira/browse/KAFKA-9255
> Project: Kafka
>  Issue Type: Bug
>  Components: documentation
>Reporter: Fábio Silva
>Assignee: Fábio Silva
>Priority: Major
>
> The documentation contains a BNF specification missing the timestamp field on 
> 'message' field entry.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (KAFKA-9255) MessageSet v1 protocol wrong specification

2019-12-02 Thread Mickael Maison (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9255?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mickael Maison resolved KAFKA-9255.
---
Fix Version/s: 2.5.0
   Resolution: Fixed

> MessageSet v1 protocol wrong specification
> --
>
> Key: KAFKA-9255
> URL: https://issues.apache.org/jira/browse/KAFKA-9255
> Project: Kafka
>  Issue Type: Bug
>  Components: documentation
>Reporter: Fábio Silva
>Assignee: Fábio Silva
>Priority: Major
> Fix For: 2.5.0
>
>
> The documentation contains a BNF specification missing the timestamp field on 
> 'message' field entry.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9255) MessageSet v1 protocol wrong specification

2019-12-02 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986099#comment-16986099
 ] 

ASF GitHub Bot commented on KAFKA-9255:
---

mimaison commented on pull request #7764: KAFKA-9255: MessageSet v1 protocol 
wrong specification
URL: https://github.com/apache/kafka/pull/7764
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> MessageSet v1 protocol wrong specification
> --
>
> Key: KAFKA-9255
> URL: https://issues.apache.org/jira/browse/KAFKA-9255
> Project: Kafka
>  Issue Type: Bug
>  Components: documentation
>Reporter: Fábio Silva
>Priority: Major
>
> The documentation contains a BNF specification missing the timestamp field on 
> 'message' field entry.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (KAFKA-9254) Topic level configuration failed

2019-12-02 Thread fenghong (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986025#comment-16986025
 ] 

fenghong edited comment on KAFKA-9254 at 12/2/19 12:43 PM:
---

We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Write data continuously during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
server.properties
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

 

 


was (Author: fenghong):
We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
server.properties
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

> Topic level configuration failed
> 
>
> Key: KAFKA-9254
> URL: https://issues.apache.org/jira/browse/KAFKA-9254
> Project: Kafka
>  Issue Type: Bug
>  Components: config, log, replication
>Affects Versions: 2.0.1
>Reporter: fenghong
>Priority: Critical
>
> We are engineers at Huobi and now encounter Kafka BUG 
> Modifying DynamicBrokerConfig more than 2 times will invalidate the topic 
> level unrelated configuration
> The bug reproduction method as follows:
>  # Set Kafka Broker config  server.properties min.insync.replicas=3
>  # Create topic test-1 and set topic‘s level config min.insync.replicas=2
>  # Dynamically modify the configuration twice as shown below
> {code:java}
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.message.timestamp.type=LogAppendTime
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.retention.ms=60480
> {code}
>  # stop a Kafka Server and found the Exception as shown below
>  org.apache.kafka.common.errors.NotEnoughReplicasException: Number of insync 
> replicas for partition test-1-0 is [2], below required minimum [3]
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (KAFKA-9254) Topic level configuration failed

2019-12-02 Thread fenghong (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986025#comment-16986025
 ] 

fenghong edited comment on KAFKA-9254 at 12/2/19 12:39 PM:
---

We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
server.properties
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 


was (Author: fenghong):
We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA

 
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
Server.propertity
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

> Topic level configuration failed
> 
>
> Key: KAFKA-9254
> URL: https://issues.apache.org/jira/browse/KAFKA-9254
> Project: Kafka
>  Issue Type: Bug
>  Components: config, log, replication
>Affects Versions: 2.0.1
>Reporter: fenghong
>Priority: Critical
>
> We are engineers at Huobi and now encounter Kafka BUG 
> Modifying DynamicBrokerConfig more than 2 times will invalidate the topic 
> level unrelated configuration
> The bug reproduction method as follows:
>  # Set Kafka Broker config  server.properties min.insync.replicas=3
>  # Create topic test-1 and set topic‘s level config min.insync.replicas=2
>  # Dynamically modify the configuration twice as shown below
> {code:java}
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.message.timestamp.type=LogAppendTime
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.retention.ms=60480
> {code}
>  # stop a Kafka Server and found the Exception as shown below
>  org.apache.kafka.common.errors.NotEnoughReplicasException: Number of insync 
> replicas for partition test-1-0 is [2], below required minimum [3]
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (KAFKA-9254) Topic level configuration failed

2019-12-02 Thread fenghong (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986025#comment-16986025
 ] 

fenghong edited comment on KAFKA-9254 at 12/2/19 12:38 PM:
---

We can reproduce this bug with the latest version (2.3.1) and this version 
(2.0.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA

 
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
Server.propertity
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 


was (Author: fenghong):
We can reproduce this bug with the latest version (2.3.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA

 
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
Server.propertity
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

> Topic level configuration failed
> 
>
> Key: KAFKA-9254
> URL: https://issues.apache.org/jira/browse/KAFKA-9254
> Project: Kafka
>  Issue Type: Bug
>  Components: config, log, replication
>Affects Versions: 2.0.1
>Reporter: fenghong
>Priority: Critical
>
> We are engineers at Huobi and now encounter Kafka BUG 
> Modifying DynamicBrokerConfig more than 2 times will invalidate the topic 
> level unrelated configuration
> The bug reproduction method as follows:
>  # Set Kafka Broker config  server.properties min.insync.replicas=3
>  # Create topic test-1 and set topic‘s level config min.insync.replicas=2
>  # Dynamically modify the configuration twice as shown below
> {code:java}
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.message.timestamp.type=LogAppendTime
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.retention.ms=60480
> {code}
>  # stop a Kafka Server and found the Exception as shown below
>  org.apache.kafka.common.errors.NotEnoughReplicasException: Number of insync 
> replicas for partition test-1-0 is [2], below required minimum [3]
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9254) Topic level configuration failed

2019-12-02 Thread fenghong (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986025#comment-16986025
 ] 

fenghong commented on KAFKA-9254:
-

We can reproduce this bug with the latest version (2.3.1)

Need to continue writing data during testing

 

OS
{code:java}
Linux  4.15.0-70-generic #79~16.04.1-Ubuntu SMP Tue Nov 12 14:01:10 UTC 2019 
x86_64 x86_64 x86_64 GNU/Linux
{code}
JAVA

 
{code:java}
java version "1.8.0_192"
Java(TM) SE Runtime Environment (build 1.8.0_192-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.192-b12, mixed mode)
{code}
Server.propertity
{code:java}
broker.id=4
delete.topic.enable=true
listeners=PLAINTEXT://xxx:9092
num.network.threads=9
num.io.threads=12
socket.send.buffer.bytes=102400
socket.receive.buffer.bytes=102400
socket.request.max.bytes=104857600
log.dirs=/data/data/kafka,/data2/data/kafka,/data3/data/kafka,/data4/data/kafka
num.partitions=1
num.recovery.threads.per.data.dir=2
offsets.topic.replication.factor=5
transaction.state.log.replication.factor=5
transaction.state.log.min.isr=3
log.retention.hours=26280
log.segment.bytes=10737418
log.retention.check.interval.ms=30
zookeeper.connect=xxx1:2181,xxx2:2181,xxx3:2181,xxx4:2181,xxx5:2181
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
group.initial.rebalance.delay.ms=500
default.replication.factor=3
min.insync.replicas=3
auto.create.topics.enable=false
unclean.leader.election.enable=false
log.cleanup.policy=delete
num.replica.fetchers=2
replica.lag.time.max.ms=1
replica.fetch.wait.max.ms=500
{code}
 

> Topic level configuration failed
> 
>
> Key: KAFKA-9254
> URL: https://issues.apache.org/jira/browse/KAFKA-9254
> Project: Kafka
>  Issue Type: Bug
>  Components: config, log, replication
>Affects Versions: 2.0.1
>Reporter: fenghong
>Priority: Critical
>
> We are engineers at Huobi and now encounter Kafka BUG 
> Modifying DynamicBrokerConfig more than 2 times will invalidate the topic 
> level unrelated configuration
> The bug reproduction method as follows:
>  # Set Kafka Broker config  server.properties min.insync.replicas=3
>  # Create topic test-1 and set topic‘s level config min.insync.replicas=2
>  # Dynamically modify the configuration twice as shown below
> {code:java}
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.message.timestamp.type=LogAppendTime
> bin/kafka-configs.sh --bootstrap-server xxx:9092 --entity-type brokers 
> --entity-default --alter --add-config log.retention.ms=60480
> {code}
>  # stop a Kafka Server and found the Exception as shown below
>  org.apache.kafka.common.errors.NotEnoughReplicasException: Number of insync 
> replicas for partition test-1-0 is [2], below required minimum [3]
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-9187) kafka.api.SaslGssapiSslEndToEndAuthorizationTest.testNoDescribeProduceOrConsumeWithoutTopicDescribeAcl

2019-12-02 Thread Bruno Cadonna (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-9187?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986016#comment-16986016
 ] 

Bruno Cadonna commented on KAFKA-9187:
--

https://builds.apache.org/job/kafka-pr-jdk11-scala2.12/9656/testReport/junit/kafka.api/SaslPlainSslEndToEndAuthorizationTest/testNoDescribeProduceOrConsumeWithoutTopicDescribeAcl/

> kafka.api.SaslGssapiSslEndToEndAuthorizationTest.testNoDescribeProduceOrConsumeWithoutTopicDescribeAcl
> --
>
> Key: KAFKA-9187
> URL: https://issues.apache.org/jira/browse/KAFKA-9187
> Project: Kafka
>  Issue Type: Test
>  Components: core
>Reporter: Bill Bejeck
>Priority: Major
>  Labels: flaky-test
>
> Failed in [https://builds.apache.org/job/kafka-pr-jdk8-scala2.11/26593/]
>  
> {noformat}
> Error Messageorg.scalatest.exceptions.TestFailedException: Consumed 0 records 
> before timeout instead of the expected 1 
> recordsStacktraceorg.scalatest.exceptions.TestFailedException: Consumed 0 
> records before timeout instead of the expected 1 records
>   at 
> org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:530)
>   at 
> org.scalatest.Assertions$.newAssertionFailedException(Assertions.scala:1389)
>   at org.scalatest.Assertions$class.fail(Assertions.scala:1091)
>   at org.scalatest.Assertions$.fail(Assertions.scala:1389)
>   at kafka.utils.TestUtils$.waitUntilTrue(TestUtils.scala:842)
>   at kafka.utils.TestUtils$.pollRecordsUntilTrue(TestUtils.scala:793)
>   at 
> kafka.utils.TestUtils$.pollUntilAtLeastNumRecords(TestUtils.scala:1334)
>   at kafka.utils.TestUtils$.consumeRecords(TestUtils.scala:1343)
>   at 
> kafka.api.EndToEndAuthorizationTest.consumeRecords(EndToEndAuthorizationTest.scala:530)
>   at 
> kafka.api.EndToEndAuthorizationTest.testNoDescribeProduceOrConsumeWithoutTopicDescribeAcl(EndToEndAuthorizationTest.scala:369)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:305)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:365)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>   at org.junit.runners.ParentRunner$4.run(ParentRunner.java:330)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:78)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:328)
>   at org.junit.runners.ParentRunner.access$100(ParentRunner.java:65)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:292)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:305)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:412)
>   at 
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110)
>   at 
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
>   at 
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38)
>   at 
> org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62)
>   at 
> org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
>   at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp

[jira] [Created] (KAFKA-9257) Allow listeners to bind on same port but different interfaces

2019-12-02 Thread Jira
Gérald Quintana created KAFKA-9257:
--

 Summary: Allow listeners to bind on same port but different 
interfaces
 Key: KAFKA-9257
 URL: https://issues.apache.org/jira/browse/KAFKA-9257
 Project: Kafka
  Issue Type: Improvement
  Components: network
Affects Versions: 2.3.1
Reporter: Gérald Quintana


It's not possible to bind two listeners to two distinct network interfaces 
using the same port
{code:java}
listeners=ONE://:9092,TWO://:9092
advertised.listeners=ONE://:9092,TWO://:9092
listener.security.protocol.map=ONE:PLAINTEXT,TWO:PLAINTEXT{code}
Kafka server returns an error:
{noformat}
[2019-12-02 13:07:18,256] ERROR Exiting Kafka due to fatal exception 
(kafka.Kafka$)
 java.lang.IllegalArgumentException: requirement failed: Each listener must 
have a different port, listeners: ONE://:9092,TWO://:9092
 at scala.Predef$.require(Predef.scala:281)
 at kafka.utils.CoreUtils$.validate$1(CoreUtils.scala:305)
 at kafka.utils.CoreUtils$.listenerListToEndPoints(CoreUtils.scala:316)
 at kafka.server.KafkaConfig.advertisedListeners(KafkaConfig.scala:1407)
 at kafka.server.KafkaConfig.validateValues(KafkaConfig.scala:1482)
 at kafka.server.KafkaConfig.(KafkaConfig.scala:1460)
 at kafka.server.KafkaConfig.(KafkaConfig.scala:1114)
 at kafka.server.KafkaConfig$.fromProps(KafkaConfig.scala:1094)
 at kafka.server.KafkaServerStartable$.fromProps(KafkaServerStartable.scala:28)
 at kafka.Kafka$.main(Kafka.scala:68)
 at kafka.Kafka.main(Kafka.scala){noformat}
 

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-6291) Cannot close EmbeddedZookeeper on Windows

2019-12-02 Thread Wahid Gazzah (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-6291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16986003#comment-16986003
 ] 

Wahid Gazzah commented on KAFKA-6291:
-

Any update on this issue please ?
I'm also using Windows os,

> Cannot close EmbeddedZookeeper on Windows
> -
>
> Key: KAFKA-6291
> URL: https://issues.apache.org/jira/browse/KAFKA-6291
> Project: Kafka
>  Issue Type: Bug
>  Components: zkclient
>Affects Versions: 0.11.0.0, 1.0.0
> Environment: Windows 10 (doesn't reproduce on Linux)
> JDK 8
>Reporter: Viliam Durina
>Priority: Major
>  Labels: windows
>
> We created {{EmbeddedZookeeper}} and {{ZkClient}} for various tests using 
> this code:
> {code:java}
> EmbeddedZookeeper zkServer = new EmbeddedZookeeper();
> ZkClient zkClient = new ZkClient("127.0.0.1:" + zkServer.port(), 
> 3, 3, ZKStringSerializer$.MODULE$);
> zkClient.close();
> zkServer.shutdown();
> {code}
> This works fine on Linux, but on Windows, the {{zkServer.shutdown()}} call 
> fails with this exception:
> {code}
> [Thread-1] ERROR org.apache.kafka.test.TestUtils - Error deleting 
> C:\Users\vilo\AppData\Local\Temp\kafka-5521621147877426083
> java.nio.file.FileSystemException: 
> C:\Users\vilo\AppData\Local\Temp\kafka-5521621147877426083\version-2\log.1: 
> The process cannot access the file because it is being used by another 
> process.
>   at 
> sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86)
>   at 
> sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97)
>   at 
> sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:102)
>   at 
> sun.nio.fs.WindowsFileSystemProvider.implDelete(WindowsFileSystemProvider.java:269)
>   at 
> sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
>   at java.nio.file.Files.delete(Files.java:1126)
>   at org.apache.kafka.common.utils.Utils$1.visitFile(Utils.java:630)
>   at org.apache.kafka.common.utils.Utils$1.visitFile(Utils.java:619)
>   at java.nio.file.Files.walkFileTree(Files.java:2670)
>   at java.nio.file.Files.walkFileTree(Files.java:2742)
>   at org.apache.kafka.common.utils.Utils.delete(Utils.java:619)
>   at org.apache.kafka.test.TestUtils$1.run(TestUtils.java:184)
> java.nio.file.FileSystemException: 
> C:\Users\vilo\AppData\Local\Temp\kafka-5521621147877426083\version-2\log.1: 
> The process cannot access the file because it is being used by another 
> process.
>   at 
> sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86)
>   at 
> sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97)
>   at 
> sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:102)
>   at 
> sun.nio.fs.WindowsFileSystemProvider.implDelete(WindowsFileSystemProvider.java:269)
>   at 
> sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
>   at java.nio.file.Files.delete(Files.java:1126)
>   at org.apache.kafka.common.utils.Utils$1.visitFile(Utils.java:630)
>   at org.apache.kafka.common.utils.Utils$1.visitFile(Utils.java:619)
>   at java.nio.file.Files.walkFileTree(Files.java:2670)
>   at java.nio.file.Files.walkFileTree(Files.java:2742)
>   at org.apache.kafka.common.utils.Utils.delete(Utils.java:619)
>   at kafka.zk.EmbeddedZookeeper.shutdown(EmbeddedZookeeper.scala:53)
>   at com.hazelcast.jet.KafkaTest.test(KafkaTest.java:32)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(

[jira] [Updated] (KAFKA-9256) mirrorMaker + kafka = java.lang.NoSuchMethodError: org.apache.kafka.common.requests.MetadataResponse.prepareResponse

2019-12-02 Thread Rudik (Jira)


 [ 
https://issues.apache.org/jira/browse/KAFKA-9256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rudik updated KAFKA-9256:
-
Description: 
kafkaBroker - 2.3.0

java - openjdk version "11.0.2" 2019-01-15

 

kafka cluster from = \{broker - os-4603:9092,os-4604:9092,os-4605:9092, auth - 
plain } \{zookeeper - os-4595:2181,os-4596:2181,os-4599:2181}

kafka cluster to = \{broker - os-4804:9092,os-4805:9092,os-4806:9092, auth - 
plain} \{zookeeper - os-4804:2181,os-4805:2181,os-4806:2181}

 

[kafka@os-4804 kafka]$ cat config/mirrormaker_consumer_pdc.properties
zookeeper.connect=os-4595:2181,os-4596:2181,os-4599:2181
bootstrap.servers=os-4603:9092,os-4604:9092,os-4605:9092
group.id=dp-MirrorMaker-group
exclude.internal.topics=true
#mirror.topics.whitelist=.*
client.id=mirror_maker_consumer

 

[kafka@os-4804 kafka]$ cat config/mirrormaker_producer.properties
zookeeper.connect=os-4804:2181,os-4805:2181,os-4806:2181
bootstrap.servers=os-4804:9092,os-4805:9092,os-4806:9092
acks=1
batch.size=100
client.id=mirror_maker_producer

 

Start MirrorMaker:

[kafka@os-4804 kafka]$ bin/kafka-mirror-maker.sh --consumer.config 
./config/mirrormaker_consumer_pdc.properties --num.streams 2 --producer.config 
./config/mirrormaker_producer.properties --whitelist="connect-configs, 
connect-offsets, connect-status, last_topic"
 WARNING: The default partition assignment strategy of the mirror maker will 
change from 'range' to 'roundrobin' in an upcoming release (so that better load 
balancing can be achiev ed). If you prefer to make this switch in advance of 
that release add the following to the corresponding config: 
'partition.assignment.strategy=org.apache.kafka.clients.consumer.R 
oundRobinAssignor'
 [2019-12-02 13:50:45,391] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.producer.ProducerConfig)
 [2019-12-02 13:50:45,483] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.consumer.ConsumerConfig)
 [2019-12-02 13:50:45,508] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.consumer.ConsumerConfig)
 [2019-12-02 13:50:45,749] WARN [Consumer clientId=dp-MirrorMaker-group-0, 
groupId=dp-MirrorMaker-group] Connection to node -2 (os-4604/10.6.107.93:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)
 [2019-12-02 13:50:45,750] WARN [Consumer clientId=dp-MirrorMaker-group-1, 
groupId=dp-MirrorMaker-group] Connection to node -2 (os-4604/10.6.107.93:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)
 [2019-12-02 13:50:45,758] WARN [Consumer clientId=dp-MirrorMaker-group-0, 
groupId=dp-MirrorMaker-group] Connection to node -3 (os-4605/10.6.107.94:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)

 

logs in KafkaBroker:

[2019-12-02 13:51:54,020] ERROR [KafkaApi-1] Error when handling request: 
clientId=dp-MirrorMaker-group-1, correlationId=269, api=METADATA, 
body=\{topics=null,allow_auto_topic_crea 
tion=true,include_cluster_authorized_operations=false,include_topic_authorized_operations=false}
 (kafka.server.KafkaApis)
 java.lang.NoSuchMethodError: 
org.apache.kafka.common.requests.MetadataResponse.prepareResponse(ILjava/util/List;Ljava/lang/String;ILjava/util/List;I)Lorg/apache/kafka/common/reque
 sts/MetadataResponse;
 at 
kafka.server.KafkaApis.$anonfun$handleTopicMetadataRequest$7(KafkaApis.scala:1103)
 at 
kafka.server.KafkaApis.$anonfun$handleTopicMetadataRequest$7$adapted(KafkaApis.scala:1096)
 at kafka.server.KafkaApis.sendResponseMaybeThrottle(KafkaApis.scala:2539)
 at kafka.server.KafkaApis.handleTopicMetadataRequest(KafkaApis.scala:1096)
 at kafka.server.KafkaApis.handle(KafkaApis.scala:116)
 at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:69)
 at java.base/java.lang.Thread.run(Thread.java:834)

 

  was:
kafkaBroker - 2.3.0

java - openjdk version "11.0.2" 2019-01-15

 

kafka cluster from = \{broker - os-4603:9092,os-4604:9092,os-4605:9092, auth - 
plain } \{zookeeper - os-4595:2181,os-4596:2181,os-4599:2181}

kafka cluster to = \{broker - os-4804:9092,os-4805:9092,os-4806:9092, auth - 
plain} \{zookeeper - os-4804:2181,os-4805:2181,os-4806:2181}

 

Start MirrorMaker:

[kafka@os-4804 kafka]$ bin/kafka-mirror-maker.sh --consumer.config 
./config/mirrormaker_consumer_pdc.properties --num.streams 2 --producer.config 
./config/mirrormaker_producer.pr operties --whitelist="connect-configs, 
connect-offsets, connect-status, last_topic"
WARNING: The default partition assignment strategy of the mirror maker will 
change from 'range' to 'roundrobin' in an upcoming release (so that better load 
balancing can be achiev ed). If you prefer to make this switch in advance of 
that release add the following to the correspond

[jira] [Created] (KAFKA-9256) mirrorMaker + kafka = java.lang.NoSuchMethodError: org.apache.kafka.common.requests.MetadataResponse.prepareResponse

2019-12-02 Thread Rudik (Jira)
Rudik created KAFKA-9256:


 Summary: mirrorMaker + kafka = java.lang.NoSuchMethodError: 
org.apache.kafka.common.requests.MetadataResponse.prepareResponse
 Key: KAFKA-9256
 URL: https://issues.apache.org/jira/browse/KAFKA-9256
 Project: Kafka
  Issue Type: Bug
  Components: mirrormaker
Affects Versions: 2.3.0
Reporter: Rudik


kafkaBroker - 2.3.0

java - openjdk version "11.0.2" 2019-01-15

 

kafka cluster from = \{broker - os-4603:9092,os-4604:9092,os-4605:9092, auth - 
plain } \{zookeeper - os-4595:2181,os-4596:2181,os-4599:2181}

kafka cluster to = \{broker - os-4804:9092,os-4805:9092,os-4806:9092, auth - 
plain} \{zookeeper - os-4804:2181,os-4805:2181,os-4806:2181}

 

Start MirrorMaker:

[kafka@os-4804 kafka]$ bin/kafka-mirror-maker.sh --consumer.config 
./config/mirrormaker_consumer_pdc.properties --num.streams 2 --producer.config 
./config/mirrormaker_producer.pr operties --whitelist="connect-configs, 
connect-offsets, connect-status, last_topic"
WARNING: The default partition assignment strategy of the mirror maker will 
change from 'range' to 'roundrobin' in an upcoming release (so that better load 
balancing can be achiev ed). If you prefer to make this switch in advance of 
that release add the following to the corresponding config: 
'partition.assignment.strategy=org.apache.kafka.clients.consumer.R 
oundRobinAssignor'
[2019-12-02 13:50:45,391] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.producer.ProducerConfig)
[2019-12-02 13:50:45,483] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.consumer.ConsumerConfig)
[2019-12-02 13:50:45,508] WARN The configuration 'zookeeper.connect' was 
supplied but isn't a known config. 
(org.apache.kafka.clients.consumer.ConsumerConfig)
[2019-12-02 13:50:45,749] WARN [Consumer clientId=dp-MirrorMaker-group-0, 
groupId=dp-MirrorMaker-group] Connection to node -2 (os-4604/10.6.107.93:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)
[2019-12-02 13:50:45,750] WARN [Consumer clientId=dp-MirrorMaker-group-1, 
groupId=dp-MirrorMaker-group] Connection to node -2 (os-4604/10.6.107.93:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)
[2019-12-02 13:50:45,758] WARN [Consumer clientId=dp-MirrorMaker-group-0, 
groupId=dp-MirrorMaker-group] Connection to node -3 (os-4605/10.6.107.94:9092) 
could not be established. Broker may not be available. 
(org.apache.kafka.clients.NetworkClient)

 

logs in KafkaBroker:

[2019-12-02 13:51:54,020] ERROR [KafkaApi-1] Error when handling request: 
clientId=dp-MirrorMaker-group-1, correlationId=269, api=METADATA, 
body=\{topics=null,allow_auto_topic_crea 
tion=true,include_cluster_authorized_operations=false,include_topic_authorized_operations=false}
 (kafka.server.KafkaApis)
java.lang.NoSuchMethodError: 
org.apache.kafka.common.requests.MetadataResponse.prepareResponse(ILjava/util/List;Ljava/lang/String;ILjava/util/List;I)Lorg/apache/kafka/common/reque
 sts/MetadataResponse;
 at 
kafka.server.KafkaApis.$anonfun$handleTopicMetadataRequest$7(KafkaApis.scala:1103)
 at 
kafka.server.KafkaApis.$anonfun$handleTopicMetadataRequest$7$adapted(KafkaApis.scala:1096)
 at kafka.server.KafkaApis.sendResponseMaybeThrottle(KafkaApis.scala:2539)
 at kafka.server.KafkaApis.handleTopicMetadataRequest(KafkaApis.scala:1096)
 at kafka.server.KafkaApis.handle(KafkaApis.scala:116)
 at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:69)
 at java.base/java.lang.Thread.run(Thread.java:834)

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (KAFKA-8264) Flaky Test PlaintextConsumerTest#testLowMaxFetchSizeForRequestAndPartition

2019-12-02 Thread Bruno Cadonna (Jira)


[ 
https://issues.apache.org/jira/browse/KAFKA-8264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16985973#comment-16985973
 ] 

Bruno Cadonna commented on KAFKA-8264:
--

{code:java}
org.scalatest.exceptions.TestFailedException: Timed out before consuming 
expected 2700 records. The number consumed was 1440.
{code}

https://builds.apache.org/job/kafka-pr-jdk11-scala2.12/9656/


> Flaky Test PlaintextConsumerTest#testLowMaxFetchSizeForRequestAndPartition
> --
>
> Key: KAFKA-8264
> URL: https://issues.apache.org/jira/browse/KAFKA-8264
> Project: Kafka
>  Issue Type: Bug
>  Components: core, unit tests
>Affects Versions: 2.0.1, 2.3.0
>Reporter: Matthias J. Sax
>Assignee: Stanislav Kozlovski
>Priority: Critical
>  Labels: flaky-test
> Fix For: 2.5.0
>
>
> [https://builds.apache.org/blue/organizations/jenkins/kafka-2.0-jdk8/detail/kafka-2.0-jdk8/252/tests]
> {quote}org.apache.kafka.common.errors.TopicExistsException: Topic 'topic3' 
> already exists.{quote}
> STDOUT
>  
> {quote}[2019-04-19 03:54:20,080] ERROR [ReplicaFetcher replicaId=1, 
> leaderId=0, fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:20,080] ERROR [ReplicaFetcher replicaId=2, leaderId=0, 
> fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:20,312] ERROR [ReplicaFetcher replicaId=2, leaderId=0, 
> fetcherId=0] Error for partition topic-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:20,313] ERROR [ReplicaFetcher replicaId=2, leaderId=1, 
> fetcherId=0] Error for partition topic-1 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:20,994] ERROR [ReplicaFetcher replicaId=1, leaderId=0, 
> fetcherId=0] Error for partition topic-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:21,727] ERROR [ReplicaFetcher replicaId=0, leaderId=2, 
> fetcherId=0] Error for partition topicWithNewMessageFormat-1 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:28,696] ERROR [ReplicaFetcher replicaId=0, leaderId=2, 
> fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:28,699] ERROR [ReplicaFetcher replicaId=1, leaderId=2, 
> fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:29,246] ERROR [ReplicaFetcher replicaId=0, leaderId=2, 
> fetcherId=0] Error for partition topic-1 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:29,247] ERROR [ReplicaFetcher replicaId=0, leaderId=1, 
> fetcherId=0] Error for partition topic-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:29,287] ERROR [ReplicaFetcher replicaId=1, leaderId=2, 
> fetcherId=0] Error for partition topic-1 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:33,408] ERROR [ReplicaFetcher replicaId=0, leaderId=2, 
> fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.server.ReplicaFetcherThread:76)
> org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server 
> does not host this topic-partition.
> [2019-04-19 03:54:33,408] ERROR [ReplicaFetcher replicaId=1, leaderId=2, 
> fetcherId=0] Error for partition __consumer_offsets-0 at offset 0 
> (kafka.s