[ 
https://issues.apache.org/jira/browse/KAFKA-7384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16607241#comment-16607241
 ] 

Vasilis Tsamis edited comment on KAFKA-7384 at 9/7/18 3:20 PM:
---------------------------------------------------------------

Our producer uses the default compression type(which is none). The exceptions 
return unknown compression type ids start from 4 to 7 which do not even exist. 
My guess is that these arbitrary ids are returned probably because the consumer 
decodes the messages in a wrong way. I guess this is why the client also throws 
the IndexOutOfBoundsException in some cases.


was (Author: vtsamis):
Our producer uses the default compression type(which is none). The exceptions 
return unknown compression type ids from 4 to 7 which do not even exist. My 
guess is that these arbitrary ids are returned probably because the consumer 
decodes the messages in a wrong way. I guess this is why the client also throws 
the IndexOutOfBoundsException in some cases.

> Compatibility issues between Kafka Brokers 1.1.0 and older kafka clients
> ------------------------------------------------------------------------
>
>                 Key: KAFKA-7384
>                 URL: https://issues.apache.org/jira/browse/KAFKA-7384
>             Project: Kafka
>          Issue Type: Bug
>          Components: consumer
>    Affects Versions: 1.1.0
>            Reporter: Vasilis Tsamis
>            Priority: Blocker
>         Attachments: logs2.txt
>
>
> Hello
> After upgrading the Kafka Brokers from 0.10.2.1 to 1.1.0, I am getting the 
> following error logs thrown by the kafka clients 0.10.2.1 & 0.10.0.1. This 
> seems to be some kind of incompatibility issue for the older clients although 
> this shouldn't be true according to the following [doc 
> 1|https://docs.confluent.io/current/installation/upgrade.html#preparation], 
> [doc2|https://cwiki-test.apache.org/confluence/display/KAFKA/KIP-35+-+Retrieving+protocol+version]
>  and 
> [thread|https://lists.apache.org/thread.html/9bc87a2c683d13fda27f01a635dba822520113cfd8fb50f3a3e82fcf@%3Cusers.kafka.apache.org%3E].
> Can someone please help on this issue? Does this mean that I have to upgrade 
> all kafka-clients to 1.1.0?
>  
> (Please also check the attached log, some extra compression type ids are also 
> occurring)
>  
> {noformat}
> java.lang.IllegalArgumentException: Unknown compression type id: 4
>         at 
> org.apache.kafka.common.record.CompressionType.forId(CompressionType.java:46)
>         at 
> org.apache.kafka.common.record.Record.compressionType(Record.java:260)
>         at 
> org.apache.kafka.common.record.LogEntry.isCompressed(LogEntry.java:89)
>         at 
> org.apache.kafka.common.record.RecordsIterator.makeNext(RecordsIterator.java:70)
>         at 
> org.apache.kafka.common.record.RecordsIterator.makeNext(RecordsIterator.java:34)
>         at 
> org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79)
>         at 
> org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45)
>         at 
> org.apache.kafka.clients.consumer.internals.Fetcher.parseCompletedFetch(Fetcher.java:785)
>         at 
> org.apache.kafka.clients.consumer.internals.Fetcher.fetchedRecords(Fetcher.java:480)
>         at 
> org.apache.kafka.clients.consumer.KafkaConsumer.pollOnce(KafkaConsumer.java:1037)
>         at 
> org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:995)
>         at 
> org.apache.camel.component.kafka.KafkaConsumer$KafkaFetchRecords.run(KafkaConsumer.java:130)
>         at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>         at java.lang.Thread.run(Thread.java:745)
> ------------------- Another kind of exception due to same reason
> java.lang.IndexOutOfBoundsException: null
>         at java.nio.Buffer.checkIndex(Buffer.java:546)
>         at java.nio.HeapByteBuffer.getInt(HeapByteBuffer.java:365) 
>         at org.apache.kafka.common.utils.Utils.sizeDelimited(Utils.java:784)
>         at org.apache.kafka.common.record.Record.value(Record.java:268)
>         at 
> org.apache.kafka.common.record.RecordsIterator$DeepRecordsIterator.<init>(RecordsIterator.java:149)
>         at 
> org.apache.kafka.common.record.RecordsIterator.makeNext(RecordsIterator.java:79)
>         at 
> org.apache.kafka.common.record.RecordsIterator.makeNext(RecordsIterator.java:34)
>         at 
> org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79)
>         at 
> org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45)
>         at 
> org.apache.kafka.clients.consumer.internals.Fetcher.parseCompletedFetch(Fetcher.java:785)
>         at 
> org.apache.kafka.clients.consumer.internals.Fetcher.fetchedRecords(Fetcher.java:480)
>         at 
> org.apache.kafka.clients.consumer.KafkaConsumer.pollOnce(KafkaConsumer.java:1037)
>         at 
> org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:995)
>         at 
> org.apache.camel.component.kafka.KafkaConsumer$KafkaFetchRecords.run(KafkaConsumer.java:130)
>         at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>         at java.lang.Thread.run(Thread.java:745)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to