Re: NPE in low level Kafka topology

2018-06-20 Thread Frank Lyaruu
p size (in case of a silent OOM exception), doesn't seem to matter I'm kinda out of ideas. On Tue, Jun 19, 2018 at 11:02 AM Frank Lyaruu wrote: > We've tried running a fresh version with yesterday morning's trunk > version, with the same result. > We're runnin

Re: NPE in low level Kafka topology

2018-06-19 Thread Frank Lyaruu
h using the > trunk version? > > > Guozhang > > On Mon, Jun 18, 2018 at 9:03 AM, Frank Lyaruu wrote: > > > Yes, here it is: > > > > https://gist.github.com/flyaruu/84b65d52a01987946b07d21bafe0d5f1 > > > > It ran completely fine for the

Re: NPE in low level Kafka topology

2018-06-18 Thread Frank Lyaruu
ted code from OneToManyGroupedProcessor ? > > Thanks > > On Mon, Jun 18, 2018 at 4:29 AM, Frank Lyaruu wrote: > > > Hi, I've upgraded our 0.11 based stream application to the trunk version, > > and I get an intermittent NPE. It's is quite a big topology, and I >

NPE in low level Kafka topology

2018-06-18 Thread Frank Lyaruu
Hi, I've upgraded our 0.11 based stream application to the trunk version, and I get an intermittent NPE. It's is quite a big topology, and I haven't succeeded in reproducing it on a simpler topology. It builds the topology, starts Kafka Streams, runs for about 20s., and then it terminates It seems

Re: Lost messages and messed up offsets

2017-11-29 Thread Frank Lyaruu
Do you commit the received messages? Either by doing it manually or setting enable.auto.commit and auto.commit.interval.ms? On Wed, Nov 29, 2017 at 11:15 PM, Tom van den Berge < tom.vandenbe...@gmail.com> wrote: > I'm using Kafka 0.10.0. > > I'm reading messages from a single topic (20 partitions

Re: NPE on startup with a low-level API based application

2017-06-15 Thread Frank Lyaruu
key and value you are trying to write? > >> > >> Thanks, > >> Apurva > >> > >> On Thu, Jun 15, 2017 at 2:28 AM, Frank Lyaruu > wrote: > >> > >>> Hey people, I see an error I haven't seen before. It is on a > lowlevel-AP

Re: NPE on startup with a low-level API based application

2017-06-15 Thread Frank Lyaruu
It seems to happen when using Streams 0.11.1 snapshot against a 0.10.2 (release) broker, the problem disappeared after I upgraded the broker. On Thu, Jun 15, 2017 at 11:28 AM, Frank Lyaruu wrote: > Hey people, I see an error I haven't seen before. It is on a lowlevel-API > ba

NPE on startup with a low-level API based application

2017-06-15 Thread Frank Lyaruu
Hey people, I see an error I haven't seen before. It is on a lowlevel-API based streams application. I've started it once, then it ran fine, then did a graceful shutdown and since then I always see this error on startup. I'm using yesterday's trunk. It seems that the MemoryRecordsBuilder overflow

Re: Losing messages in Kafka Streams after upgrading

2017-06-07 Thread Frank Lyaruu
should happen. Not sure why you don't observe this. And thus, > the producer should use this timestamp to write the records. > > How did you verify the timestamps that are set for your output records? > > > -Matthias > > > On 6/5/17 6:15 AM, Frank Lyaruu wrote: >

Re: Losing messages in Kafka Streams after upgrading

2017-06-05 Thread Frank Lyaruu
stamp semantics. We need to revisit on timestamps can be maintained > across the topology in Streams. > > Guozhang > > On Sat, Jun 3, 2017 at 10:54 AM, Frank Lyaruu wrote: > > > Hi Matthias, > > > > Ok, that clarifies quite a bit. I never really went into

Re: Losing messages in Kafka Streams after upgrading

2017-06-03 Thread Frank Lyaruu
implement a custom timestamp extractor and adjust the > timestamps accordingly ("stream time" is based on whatever timestamp > extractor return) > > However, if you have records with old timestamps, I am wondering why > they are not truncated in your input topic? Do you no

Losing messages in Kafka Streams after upgrading

2017-06-02 Thread Frank Lyaruu
Hi Kafka people, I'm running an application that pushes database changes into a Kafka topic. I'm also running a Kafka streams application that listens to these topics, and groups them using the high level API, and inserts them to another database. All topics are compacted, with the exception of t

Broker does not seem to honor min.compaction.lag.ms

2017-05-29 Thread Frank Lyaruu
Hi Kafka people, I've configured my brokers to use one hour (360 ms) for the min.compaction.lag.ms. In the logs I see that this is picked up in the 'created log' line. I add a bit over a million messages, and I see the new log segments appearing ('Rolling new log segment') but after a few min

Re: Creating read-only state stores from a compacted topic

2017-05-17 Thread Frank Lyaruu
) > > .build(); > > > > > > Does that help? > > > > Thanks, > > Damian > > > > > > > > On Wed, 17 May 2017 at 06:09 Frank Lyaruu wrote: > > > >> Hi Kafka people, > >> > >> I'm using the low

Creating read-only state stores from a compacted topic

2017-05-16 Thread Frank Lyaruu
Hi Kafka people, I'm using the low level API that often creates a simple state store based on a (single partition, compacted) topic (It feeds all messages and simply stores all those messages as-is). Now all these stores get their own 'changelog' topic, to back the state store. This is fine, but

Repartition topics in Kafka Streams when there is only one partition

2017-05-15 Thread Frank Lyaruu
Hi, I'm using quite a lot of topics, and the speed on each topic isn't extreme, so I generally use a single partition. In this situation, the high level API of Kafka Streams still makes repartition topics, but I don't see there use. Wouldn't it make sense to skip repartition topics in this case,

Re: Kafka Streams reports: "The timestamp of the message is out of acceptable range"

2017-05-15 Thread Frank Lyaruu
t; of this JIRA I believe: https://issues.apache.org/jira/browse/KAFKA-4861 < > https://issues.apache.org/jira/browse/KAFKA-4861> > > Thanks > Eno > > On 14 May 2017, at 23:09, Frank Lyaruu wrote: > > > > Hi Kafka people... > > > > After a bit of tuning a

Kafka Streams reports: "The timestamp of the message is out of acceptable range"

2017-05-14 Thread Frank Lyaruu
Hi Kafka people... After a bit of tuning and an upgrade to Kafka 0.10.1.2, this error starts showing up and the whole thing kind of dies. 2017-05-14 18:51:52,342 | ERROR | hread-3-producer | RecordCollectorImpl | 91 - com.dexels.kafka.streams - 0.0.115.201705131415 | task [1_0] Error s

Re: Commits of slow moving topics in Kafka Streams time out

2017-03-22 Thread Frank Lyaruu
ncrease the retention minutes to something large? > > Thanks > Eno > > On 21 Mar 2017, at 19:13, Frank Lyaruu wrote: > > > > Hi Kafka people, > > > > We have a Kafka Streams application that replicates a database, and > > transforms it to a different data model.

Commits of slow moving topics in Kafka Streams time out

2017-03-21 Thread Frank Lyaruu
Hi Kafka people, We have a Kafka Streams application that replicates a database, and transforms it to a different data model. Some tables/topics move fast, with many changes a second, some might be dormant for months. For those slow moving topics, we have some trouble with the 'offsets.retention.m

Re: Implementing a non-key in Kafka Streams using the Processor API

2017-02-21 Thread Frank Lyaruu
gt; design is particularly challenging and why Global KTables was chosen > instead. I'm not sure if you still want to pursue that original design, > since it is not proven to work. > > Guozhang, perhaps we need to add a note saying that Global KTables is the > new design?

Implementing a non-key in Kafka Streams using the Processor API

2017-02-20 Thread Frank Lyaruu
Hi all, I'm trying to implement joining two Kafka tables using a 'remote' key, basically as described here: https://cwiki.apache.org/confluence/display/KAFKA/Discussion%3A+Non-key+KTable-KTable+Joins Under the "Implementation Details" there is one line I don't know how to do: 1. First of al

Re: RocksDB error when running Kafka with a larger data set

2017-02-15 Thread Frank Lyaruu
most > informative! > Is there anything obvious in the RocksDB LOG? > > Thanks, > Damian > > > On Wed, 15 Feb 2017 at 06:38 Frank Lyaruu wrote: > > > Hi Kafka crew, > > > > I'm rolling out a Kafka Streams application on a pretty large dataset, > a

RocksDB error when running Kafka with a larger data set

2017-02-15 Thread Frank Lyaruu
Hi Kafka crew, I'm rolling out a Kafka Streams application on a pretty large dataset, and I see some exceptions that worry me, which I haven't seen with smaller data sets (Using Kafka streams trunk of a few days ago) 2017-02-15 15:25:14,431 | INFO | StreamThread-42 | StreamTask | 86 -

Re: Error in Kafka Streams (that shouldn't be there)

2016-12-21 Thread Frank Lyaruu
ted transition so the first error is the cause > probably. However, I can't tell what happened before since a normal > shutdown also looks the same. > > Eno > > On 16 Dec 2016, at 11:35, Frank Lyaruu wrote: > > > > Hi people, > > > > I'm run

Error in Kafka Streams (that shouldn't be there)

2016-12-16 Thread Frank Lyaruu
Hi people, I'm running a Kafka Streams app (trunk build from a few days ago), I see this error from time to time: Exception in thread "StreamThread-18" java.lang.NullPointerException at org.apache.kafka.streams.processor.internals.StreamThread$5. apply(StreamThread.java:998) at org.apache.kafka.

Kafka Connect gets into a rebalance loop

2016-12-16 Thread Frank Lyaruu
Hi people, I've just deployed my Kafka Streams / Connect (I only use a connect sink to mongodb) application on a cluster of four instances (4 containers on 2 machines) and now it seems to get into a sort of rebalancing loop, and I don't get much in mongodb, I've got a little bit of data at the beg

Re: Initializing StateStores takes *really* long for large datasets

2016-11-28 Thread Frank Lyaruu
the RocksDB memory settings, yes the off heap memory usage does > > sneak under the radar. There is a memory management story for Kafka > Streams > > that is yet to be started. This would involve limiting the off-heap > memory > > that RocksDB uses. > > > > Thanks, &g

Re: Initializing StateStores takes *really* long for large datasets

2016-11-25 Thread Frank Lyaruu
default in streams is 3 > > However, i'm no expert on RocksDB and i suggest you have look at > https://github.com/facebook/rocksdb/wiki/RocksDB-Tuning-Guide for more > info. > > Thanks, > Damian > > On Fri, 25 Nov 2016 at 13:02 Frank Lyaruu wrote: > > > @

Re: Initializing StateStores takes *really* long for large datasets

2016-11-25 Thread Frank Lyaruu
son wrote: > What kind of disk are you using for the rocksdb store? ie spinning or ssd? > > 2016-11-25 12:51 GMT+01:00 Damian Guy : > > > Hi Frank, > > > > Is this on a restart of the application? > > > > Thanks, > > Damian > > > >

Initializing StateStores takes *really* long for large datasets

2016-11-25 Thread Frank Lyaruu
Hi y'all, I have a reasonably simple KafkaStream application, which merges about 20 topics a few times. The thing is, some of those topic datasets are pretty big, about 10M messages. In total I've got about 200Gb worth of state in RocksDB, the largest topic is 38 Gb. I had set the MAX_POLL_INTERV

Kafka Streams fails permanently when used with an unstable network

2016-10-30 Thread Frank Lyaruu
I have a remote Kafka cluster, to which I connect using a VPN and a not-so-great WiFi network. That means that sometimes the Kafka Client loses briefly loses connectivity. When it regains a connection after a while, I see: org.apache.kafka.clients.consumer.CommitFailedException: Commit cannot be c

Re: Occasional NPE in NamedCache

2016-10-18 Thread Frank Lyaruu
ing the bug, and many thanks to Damian for the > > quick catch! > > > > On Thu, Oct 13, 2016 at 12:30 PM, Frank Lyaruu > wrote: > > > >> The issue seems to be gone. Amazing work, thanks...! > >> > >> On Thu, Oct 13, 2016 at 6:56 PM, Damian Guy >

Re: Occasional NPE in NamedCache

2016-10-13 Thread Frank Lyaruu
2016 at 17:46 Damian Guy wrote: > > > Hi Frank, > > > > Thanks for reporting. Can you provide a sample of the join you are > > running? > > > > Thanks, > > Damian > > > > On Thu, 13 Oct 2016 at 16:10 Frank Lyaruu wrote: > > > >

Occasional NPE in NamedCache

2016-10-13 Thread Frank Lyaruu
Hi Kafka people, I'm joining a bunch of Kafka Topics using Kafka Streams, with the Kafka 0.10.1 release candidate. It runs ok for a few thousand of messages, and then it dies with the following exception: Exception in thread "StreamThread-1" java.lang.NullPointerException at org.apache.kafka.str