2 things to check: 1. Are you connecting to Kafka 0.10 broker? 2. Which port are you using? Recent Kafka clients must point to the kafka broker port directly. Older clients were connecting through a zookeeper and used different host/port.
Andrew On Sun, Oct 30, 2016, 1:11 PM Daniel Einspanjer < [email protected]> wrote: > These db rows are fairly small. Five or six small text fields and five or > six integer fields plus a timestamp. Looking at the queue in NiFi, about > 440 bytes each. > > > If I exported the template correctly, the flow should show that I am > trying to generate a set of query statement flow files, each 10k records, > and then executesql on them. Next I use SplitAvro and AvroToJSON to get a > set of flow files, each one record in json format. Those are what I'm > feeding in to Kafka. > > -Daniel > > On Oct 30, 2016 11:00 AM, "Joe Witt" <[email protected]> wrote: > > Daniel > > How large is each object you are trying to write to Kafka? > > Since it was working based on a different source of data but now > problematic that is the direction I am looking in in terms of changes. > Output of the db stuff could need demarcation for example. > > >
