cool.

On Tue, Mar 17, 2015 at 10:15 AM, Zakee <kzak...@netzero.net> wrote:

> Hi Mayuresh,
>
> The logs are already attached and are in reverse order starting backwards
> from [2015-03-14 07:46:52,517] to the time when brokers were started.
>
> Thanks
> Zakee
>
>
>
> > On Mar 17, 2015, at 12:07 AM, Mayuresh Gharat <
> gharatmayures...@gmail.com> wrote:
> >
> > Hi Zakee,
> >
> > Thanks for the logs. Can you paste earlier logs from broker-3 up to :
> >
> > [2015-03-14 07:46:52,517] ERROR [ReplicaFetcherThread-2-4], Current
> > offset 1754769769 for partition [Topic22kv,5] out of range; reset
> > offset to 1400864851 (kafka.server.ReplicaFetcherThread)
> >
> > That would help us figure out what was happening on this broker before it
> > issued a replicaFetch request to broker-4.
> >
> > Thanks,
> >
> > Mayuresh
> >
> > On Mon, Mar 16, 2015 at 11:32 PM, Zakee <kzak...@netzero.net> wrote:
> >
> >> Hi Mayuresh,
> >>
> >> Here are the logs.
> >>
> >> ____________________________________________________________
> >> Old School Yearbook Pics
> >> View Class Yearbooks Online Free. Search by School & Year. Look Now!
> >>
> http://thirdpartyoffers.netzero.net/TGL3231/5507ca8137dc94a805e6bst01vuc
> >>
> >>
> >> Thanks,
> >> Kazim Zakee
> >>
> >>
> >>
> >>> On Mar 16, 2015, at 10:48 AM, Mayuresh Gharat <
> >> gharatmayures...@gmail.com> wrote:
> >>>
> >>> Can you provide more logs (complete) on Broker 3 till time :
> >>>
> >>> *[2015-03-14 07:46:52,517*] WARN [ReplicaFetcherThread-2-4], Replica 3
> >> for
> >>> partition [Topic22kv,5] reset its fetch offset from 1400864851 to
> current
> >>> leader 4's start offset 1400864851 (kafka.server.ReplicaFetcherThread)
> >>>
> >>> I would like to see logs from time much before it sent the fetch
> request
> >> to
> >>> Broker 4 to the time above. I want to check if in any case Broker 3
> was a
> >>> leader before broker 4 took over.
> >>>
> >>> Additional logs will help.
> >>>
> >>>
> >>> Thanks,
> >>>
> >>> Mayuresh
> >>>
> >>>
> >>>
> >>> On Sat, Mar 14, 2015 at 8:35 PM, Zakee <kzak...@netzero.net> wrote:
> >>>
> >>>> log.cleanup.policy is delete not compact.
> >>>> log.cleaner.enable=true
> >>>> log.cleaner.threads=5
> >>>> log.cleanup.policy=delete
> >>>> log.flush.scheduler.interval.ms=3000
> >>>> log.retention.minutes=1440
> >>>> log.segment.bytes=1073741824  (1gb)
> >>>>
> >>>> Messages are keyed but not compressed, producer async and uses kafka
> >>>> default partitioner.
> >>>> String message = msg.getString();
> >>>> String uniqKey = ""+rnd.nextInt();// random key
> >>>> String partKey = getPartitionKey();// partition key
> >>>> KeyedMessage<String, String> data = new KeyedMessage<String,
> >>>> String>(this.topicName, uniqKey, partKey, message);
> >>>> producer.send(data);
> >>>>
> >>>> Thanks
> >>>> Zakee
> >>>>
> >>>>
> >>>>
> >>>>> On Mar 14, 2015, at 4:23 PM, gharatmayures...@gmail.com wrote:
> >>>>>
> >>>>> Is your topic log compacted? Also if it is are the messages keyed? Or
> >>>> are the messages compressed?
> >>>>>
> >>>>> Thanks,
> >>>>>
> >>>>> Mayuresh
> >>>>>
> >>>>> Sent from my iPhone
> >>>>>
> >>>>>> On Mar 14, 2015, at 2:02 PM, Zakee <kzak...@netzero.net <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>
> >>>>>> Thanks, Jiangjie for helping resolve the kafka controller migration
> >>>> driven partition leader rebalance issue. The logs are much cleaner
> now.
> >>>>>>
> >>>>>> There are a few incidences of Out of range offset even though  there
> >> is
> >>>> no consumers running, only producers and replica fetchers. I was
> trying
> >> to
> >>>> relate to a cause, looks like compaction (log segment deletion)
> causing
> >>>> this. Not sure whether this is expected behavior.
> >>>>>>
> >>>>>> Broker-4:
> >>>>>> [2015-03-14 07:46:52,338] ERROR [Replica Manager on Broker 4]: Error
> >>>> when processing fetch request for partition [Topic22kv,5] offset
> >> 1754769769
> >>>> from follower with correlation id 1645671. Possible cause: Request for
> >>>> offset 1754769769 but we only have log segments in the range
> 1400864851
> >> to
> >>>> 1754769732. (kafka.server.ReplicaManager)
> >>>>>>
> >>>>>> Broker-3:
> >>>>>> [2015-03-14 07:46:52,356] INFO The cleaning for partition
> >> [Topic22kv,5]
> >>>> is aborted and paused (kafka.log.LogCleaner)
> >>>>>> [2015-03-14 07:46:52,408] INFO Scheduling log segment 1400864851 for
> >>>> log Topic22kv-5 for deletion. (kafka.log.Log)
> >>>>>> …
> >>>>>> [2015-03-14 07:46:52,421] INFO Compaction for partition
> [Topic22kv,5]
> >>>> is resumed (kafka.log.LogCleaner)
> >>>>>> [2015-03-14 07:46:52,517] ERROR [ReplicaFetcherThread-2-4], Current
> >>>> offset 1754769769 for partition [Topic22kv,5] out of range; reset
> >> offset to
> >>>> 1400864851 (kafka.server.ReplicaFetcherThread)
> >>>>>> [2015-03-14 07:46:52,517] WARN [ReplicaFetcherThread-2-4], Replica 3
> >>>> for partition [Topic22kv,5] reset its fetch offset from 1400864851 to
> >>>> current leader 4's start offset 1400864851
> >>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>
> >>>>>> ____________________________________________________________
> >>>>>> Old School Yearbook Pics
> >>>>>> View Class Yearbooks Online Free. Search by School & Year. Look Now!
> >>>>>>
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3231/5504a2032e49422021991st02vuc <
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3231/5504a2032e49422021991st02vuc>
> >>>>>> <topic22kv_746a_314_logs.txt>
> >>>>>>
> >>>>>>
> >>>>>> Thanks
> >>>>>> Zakee
> >>>>>>
> >>>>>>> On Mar 9, 2015, at 12:18 PM, Zakee <kzak...@netzero.net> wrote:
> >>>>>>>
> >>>>>>> No broker restarts.
> >>>>>>>
> >>>>>>> Created a kafka issue:
> >>>> https://issues.apache.org/jira/browse/KAFKA-2011 <
> >>>> https://issues.apache.org/jira/browse/KAFKA-2011>
> >>>>>>>
> >>>>>>>>> Logs for rebalance:
> >>>>>>>>> [2015-03-07 16:52:48,969] INFO [Controller 2]: Resuming preferred
> >>>> replica election for partitions: (kafka.controller.KafkaController)
> >>>>>>>>> [2015-03-07 16:52:48,969] INFO [Controller 2]: Partitions that
> >>>> completed preferred replica election:
> (kafka.controller.KafkaController)
> >>>>>>>>> …
> >>>>>>>>> [2015-03-07 12:07:06,783] INFO [Controller 4]: Resuming preferred
> >>>> replica election for partitions: (kafka.controller.KafkaController)
> >>>>>>>>> ...
> >>>>>>>>> [2015-03-07 09:10:41,850] INFO [Controller 3]: Resuming preferred
> >>>> replica election for partitions: (kafka.controller.KafkaController)
> >>>>>>>>> ...
> >>>>>>>>> [2015-03-07 08:26:56,396] INFO [Controller 1]: Starting preferred
> >>>> replica leader election for partitions
> >> (kafka.controller.KafkaController)
> >>>>>>>>> ...
> >>>>>>>>> [2015-03-06 16:52:59,506] INFO [Controller 2]: Partitions
> >> undergoing
> >>>> preferred replica election:  (kafka.controller.KafkaController)
> >>>>>>>>>
> >>>>>>>>> Also, I still see lots of below errors (~69k) going on in the
> logs
> >>>> since the restart. Is there any other reason than rebalance for these
> >>>> errors?
> >>>>>>>>>
> >>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-2-5], Error
> >>>> for partition [Topic-11,7] to broker 5:class
> >>>> kafka.common.NotLeaderForPartitionException
> >>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-1-5], Error
> >>>> for partition [Topic-2,25] to broker 5:class
> >>>> kafka.common.NotLeaderForPartitionException
> >>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-2-5], Error
> >>>> for partition [Topic-2,21] to broker 5:class
> >>>> kafka.common.NotLeaderForPartitionException
> >>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-1-5], Error
> >>>> for partition [Topic-22,9] to broker 5:class
> >>>> kafka.common.NotLeaderForPartitionException
> >>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>
> >>>>>>>
> >>>>>>>> Could you paste the related logs in controller.log?
> >>>>>>> What specifically should I search for in the logs?
> >>>>>>>
> >>>>>>> Thanks,
> >>>>>>> Zakee
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>> On Mar 9, 2015, at 11:35 AM, Jiangjie Qin
> <j...@linkedin.com.INVALID
> >>>> <mailto:j...@linkedin.com.INVALID>> wrote:
> >>>>>>>>
> >>>>>>>> Is there anything wrong with brokers around that time? E.g. Broker
> >>>> restart?
> >>>>>>>> The log you pasted are actually from replica fetchers. Could you
> >>>> paste the
> >>>>>>>> related logs in controller.log?
> >>>>>>>>
> >>>>>>>> Thanks.
> >>>>>>>>
> >>>>>>>> Jiangjie (Becket) Qin
> >>>>>>>>
> >>>>>>>>> On 3/9/15, 10:32 AM, "Zakee" <kzak...@netzero.net <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>>>>
> >>>>>>>>> Correction: Actually  the rebalance happened quite until 24 hours
> >>>> after
> >>>>>>>>> the start, and thats where below errors were found. Ideally
> >> rebalance
> >>>>>>>>> should not have happened at all.
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Thanks
> >>>>>>>>> Zakee
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>>> On Mar 9, 2015, at 10:28 AM, Zakee <kzak...@netzero.net
> <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> Hmm, that sounds like a bug. Can you paste the log of leader
> >>>> rebalance
> >>>>>>>>>>> here?
> >>>>>>>>>> Thanks for you suggestions.
> >>>>>>>>>> It looks like the rebalance actually happened only once soon
> >> after I
> >>>>>>>>>> started with clean cluster and data was pushed, it didn’t happen
> >>>> again
> >>>>>>>>>> so far, and I see the partitions leader counts on brokers did
> not
> >>>> change
> >>>>>>>>>> since then. One of the brokers was constantly showing 0 for
> >>>> partition
> >>>>>>>>>> leader count. Is that normal?
> >>>>>>>>>>
> >>>>>>>>>> Also, I still see lots of below errors (~69k) going on in the
> logs
> >>>>>>>>>> since the restart. Is there any other reason than rebalance for
> >>>> these
> >>>>>>>>>> errors?
> >>>>>>>>>>
> >>>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-2-5],
> Error
> >>>> for
> >>>>>>>>>> partition [Topic-11,7] to broker 5:class
> >>>>>>>>>> kafka.common.NotLeaderForPartitionException
> >>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-1-5],
> Error
> >>>> for
> >>>>>>>>>> partition [Topic-2,25] to broker 5:class
> >>>>>>>>>> kafka.common.NotLeaderForPartitionException
> >>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-2-5],
> Error
> >>>> for
> >>>>>>>>>> partition [Topic-2,21] to broker 5:class
> >>>>>>>>>> kafka.common.NotLeaderForPartitionException
> >>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>> [2015-03-07 14:23:28,963] ERROR [ReplicaFetcherThread-1-5],
> Error
> >>>> for
> >>>>>>>>>> partition [Topic-22,9] to broker 5:class
> >>>>>>>>>> kafka.common.NotLeaderForPartitionException
> >>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>>
> >>>>>>>>>>> Some other things to check are:
> >>>>>>>>>>> 1. The actual property name is auto.leader.rebalance.enable,
> not
> >>>>>>>>>>> auto.leader.rebalance. You’ve probably known this, just to
> double
> >>>>>>>>>>> confirm.
> >>>>>>>>>> Yes
> >>>>>>>>>>
> >>>>>>>>>>> 2. In zookeeper path, can you verify
> >>>> /admin/preferred_replica_election
> >>>>>>>>>>> does not exist?
> >>>>>>>>>> ls /admin
> >>>>>>>>>> [delete_topics]
> >>>>>>>>>> ls /admin/preferred_replica_election
> >>>>>>>>>> Node does not exist: /admin/preferred_replica_election
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> Thanks
> >>>>>>>>>> Zakee
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>> On Mar 7, 2015, at 10:49 PM, Jiangjie Qin
> >>>> <j...@linkedin.com.INVALID <mailto:j...@linkedin.com.INVALID>>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> Hmm, that sounds like a bug. Can you paste the log of leader
> >>>> rebalance
> >>>>>>>>>>> here?
> >>>>>>>>>>> Some other things to check are:
> >>>>>>>>>>> 1. The actual property name is auto.leader.rebalance.enable,
> not
> >>>>>>>>>>> auto.leader.rebalance. You’ve probably known this, just to
> double
> >>>>>>>>>>> confirm.
> >>>>>>>>>>> 2. In zookeeper path, can you verify
> >>>> /admin/preferred_replica_election
> >>>>>>>>>>> does not exist?
> >>>>>>>>>>>
> >>>>>>>>>>> Jiangjie (Becket) Qin
> >>>>>>>>>>>
> >>>>>>>>>>>> On 3/7/15, 10:24 PM, "Zakee" <kzak...@netzero.net <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>> I started with  clean cluster and started to push data. It
> still
> >>>> does
> >>>>>>>>>>>> the
> >>>>>>>>>>>> rebalance at random durations even though the
> >>>> auto.leader.relabalance
> >>>>>>>>>>>> is
> >>>>>>>>>>>> set to false.
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks
> >>>>>>>>>>>> Zakee
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>> On Mar 6, 2015, at 3:51 PM, Jiangjie Qin
> >>>> <j...@linkedin.com.INVALID <mailto:j...@linkedin.com.INVALID>>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Yes, the rebalance should not happen in that case. That is a
> >>>> little
> >>>>>>>>>>>>> bit
> >>>>>>>>>>>>> strange. Could you try to launch a clean Kafka cluster with
> >>>>>>>>>>>>> auto.leader.election disabled and try push data?
> >>>>>>>>>>>>> When leader migration occurs, NotLeaderForPartition exception
> >> is
> >>>>>>>>>>>>> expected.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Jiangjie (Becket) Qin
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>> On 3/6/15, 3:14 PM, "Zakee" <kzak...@netzero.net <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Yes, Jiangjie, I do see lots of these errors "Starting
> >> preferred
> >>>>>>>>>>>>>> replica
> >>>>>>>>>>>>>> leader election for partitions” in logs. I also see lot of
> >>>> Produce
> >>>>>>>>>>>>>> request failure warnings in with the NotLeader Exception.
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> I tried switching off the auto.leader.relabalance to false.
> I
> >> am
> >>>>>>>>>>>>>> still
> >>>>>>>>>>>>>> noticing the rebalance happening. My understanding was the
> >>>> rebalance
> >>>>>>>>>>>>>> will
> >>>>>>>>>>>>>> not happen when this is set to false.
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Thanks
> >>>>>>>>>>>>>> Zakee
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> On Feb 25, 2015, at 5:17 PM, Jiangjie Qin
> >>>>>>>>>>>>>>> <j...@linkedin.com.INVALID <mailto:
> j...@linkedin.com.INVALID
> >>>>
> >>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> I don’t think num.replica.fetchers will help in this case.
> >>>>>>>>>>>>>>> Increasing
> >>>>>>>>>>>>>>> number of fetcher threads will only help in cases where you
> >>>> have a
> >>>>>>>>>>>>>>> large
> >>>>>>>>>>>>>>> amount of data coming into a broker and more replica
> fetcher
> >>>>>>>>>>>>>>> threads
> >>>>>>>>>>>>>>> will
> >>>>>>>>>>>>>>> help keep up. We usually only use 1-2 for each broker. But
> in
> >>>> your
> >>>>>>>>>>>>>>> case,
> >>>>>>>>>>>>>>> it looks that leader migration cause issue.
> >>>>>>>>>>>>>>> Do you see anything else in the log? Like preferred leader
> >>>>>>>>>>>>>>> election?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Jiangjie (Becket) Qin
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> On 2/25/15, 5:02 PM, "Zakee" <kzak...@netzero.net <mailto:
> >>>> kzak...@netzero.net>
> >>>>>>>>>>>>>>> <mailto:kzak...@netzero.net <mailto:kzak...@netzero.net>>>
> >>>> wrote:
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Thanks, Jiangjie.
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Yes, I do see under partitions usually shooting every
> hour.
> >>>>>>>>>>>>>>>> Anythings
> >>>>>>>>>>>>>>>> that
> >>>>>>>>>>>>>>>> I could try to reduce it?
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> How does "num.replica.fetchers" affect the replica sync?
> >>>> Currently
> >>>>>>>>>>>>>>>> have
> >>>>>>>>>>>>>>>> configured 7 each of 5 brokers.
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> -Zakee
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> On Wed, Feb 25, 2015 at 4:17 PM, Jiangjie Qin
> >>>>>>>>>>>>>>>> <j...@linkedin.com.invalid <mailto:
> >> j...@linkedin.com.invalid
> >>>>>>
> >>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> These messages are usually caused by leader migration. I
> >>>> think as
> >>>>>>>>>>>>>>>>> long
> >>>>>>>>>>>>>>>>> as
> >>>>>>>>>>>>>>>>> you don¹t see this lasting for ever and got a bunch of
> >> under
> >>>>>>>>>>>>>>>>> replicated
> >>>>>>>>>>>>>>>>> partitions, it should be fine.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Jiangjie (Becket) Qin
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> On 2/25/15, 4:07 PM, "Zakee" <kzak...@netzero.net
> >> <mailto:
> >>>> kzak...@netzero.net>> wrote:
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> Need to know if I should I be worried about this or
> ignore
> >>>> them.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> I see tons of these exceptions/warnings in the broker
> >> logs,
> >>>> not
> >>>>>>>>>>>>>>>>>> sure
> >>>>>>>>>>>>>>>>> what
> >>>>>>>>>>>>>>>>>> causes them and what could be done to fix them.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> ERROR [ReplicaFetcherThread-3-5], Error for partition
> >>>>>>>>>>>>>>>>>> [TestTopic]
> >>>>>>>>>>>>>>>>>> to
> >>>>>>>>>>>>>>>>>> broker
> >>>>>>>>>>>>>>>>>> 5:class kafka.common.NotLeaderForPartitionException
> >>>>>>>>>>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>>>>>>>>>> [2015-02-25 11:01:41,785] ERROR
> >> [ReplicaFetcherThread-3-5],
> >>>>>>>>>>>>>>>>>> Error
> >>>>>>>>>>>>>>>>>> for
> >>>>>>>>>>>>>>>>>> partition [TestTopic] to broker 5:class
> >>>>>>>>>>>>>>>>>> kafka.common.NotLeaderForPartitionException
> >>>>>>>>>>>>>>>>>> (kafka.server.ReplicaFetcherThread)
> >>>>>>>>>>>>>>>>>> [2015-02-25 11:01:41,785] WARN [Replica Manager on
> Broker
> >>>> 2]:
> >>>>>>>>>>>>>>>>>> Fetch
> >>>>>>>>>>>>>>>>>> request
> >>>>>>>>>>>>>>>>>> with correlation id 950084 from client
> >>>> ReplicaFetcherThread-1-2
> >>>>>>>>>>>>>>>>>> on
> >>>>>>>>>>>>>>>>>> partition [TestTopic,2] failed due to Leader not local
> for
> >>>>>>>>>>>>>>>>>> partition
> >>>>>>>>>>>>>>>>>> [TestTopic,2] on broker 2 (kafka.server.ReplicaManager)
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> Any ideas?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> -Zakee
> >>>>>>>>>>>>>>>>>>
> >> ____________________________________________________________
> >>>>>>>>>>>>>>>>>> Next Apple Sensation
> >>>>>>>>>>>>>>>>>> 1 little-known path to big profits
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>> http://thirdpartyoffers.netzero.net/TGL3231/54ee63b9e704b63b94061 <
> >>>> http://thirdpartyoffers.netzero.net/TGL3231/54ee63b9e704b63b94061>
> >>>>>>>>>>>>>>>>>> st0
> >>>>>>>>>>>>>>>>>> 3v
> >>>>>>>>>>>>>>>>>> uc
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >> ____________________________________________________________
> >>>>>>>>>>>>>>>>> Extended Stay America
> >>>>>>>>>>>>>>>>> Get Fantastic Amenities, low rates! Kitchen, Ample
> >> Workspace,
> >>>>>>>>>>>>>>>>> Free
> >>>>>>>>>>>>>>>>> WIFI
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>> http://thirdpartyoffers.netzero.net/TGL3255/54ee66f26da6f66f10ad4m <
> >>>> http://thirdpartyoffers.netzero.net/TGL3255/54ee66f26da6f66f10ad4m>
> >>>>>>>>>>>>>>>>> p02
> >>>>>>>>>>>>>>>>> du
> >>>>>>>>>>>>>>>>> c
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>
> ____________________________________________________________
> >>>>>>>>>>>>>>> Extended Stay America
> >>>>>>>>>>>>>>> Official Site. Free WIFI, Kitchens. Our best rates here,
> >>>>>>>>>>>>>>> guaranteed.
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>
> >>>> http://thirdpartyoffers.netzero.net/TGL3255/54ee80744cfa7747461mp13d
> <
> >>>> http://thirdpartyoffers.netzero.net/TGL3255/54ee80744cfa7747461mp13d>
> >>>>>>>>>>>>>>> uc
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> <
> >>>> http://thirdpartyoffers.netzero.net/TGL3255/54ee80744cfa7747461mp13
> >>>>>>>>>>>>>>> duc
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> ____________________________________________________________
> >>>>>>>>>>>>> The WORST exercise for aging
> >>>>>>>>>>>>> Avoid this &#34;healthy&#34; exercise to look & feel 5-10
> years
> >>>>>>>>>>>>> YOUNGER
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>
> http://thirdpartyoffers.netzero.net/TGL3255/54fa40e98a0e640e81196mp07d
> >> <
> >>>>
> http://thirdpartyoffers.netzero.net/TGL3255/54fa40e98a0e640e81196mp07d>
> >>>>>>>>>>>>> uc
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> ____________________________________________________________
> >>>>>>>>>>> Seabourn Luxury Cruises
> >>>>>>>>>>> Receive special offers from the World&#39;s Finest Small-Ship
> >>>> Cruise
> >>>>>>>>>>> Line!
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/54fbf3b0f058073b02901mp14duc <
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/54fbf3b0f058073b02901mp14duc>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> ____________________________________________________________
> >>>>>>>> Discover Seabourn
> >>>>>>>> A journey as beautiful as the destination, request a brochure
> today!
> >>>>>>>>
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/54fdebfe6a2a36bfb0bb3mp10duc <
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/54fdebfe6a2a36bfb0bb3mp10duc>
> >>>>>>>
> >>>>>>>
> >>>>>>> Thanks
> >>>>>>> Zakee
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> ____________________________________________________________
> >>>>>>> Want to place your ad here?
> >>>>>>> Advertise on United Online
> >>>>>>>
> >>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/54fdf80bc575a780b0397mp05duc
> >>>>>>
> >>>>> ____________________________________________________________
> >>>>> What's your flood risk?
> >>>>> Find flood maps, interactive tools, FAQs, and agents in your area.
> >>>>>
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/5504cccfca43a4ccf0a56mp08duc
> >>>> <
> >>
> http://thirdpartyoffers.netzero.net/TGL3255/5504cccfca43a4ccf0a56mp08duc>
> >>>>
> >>>
> >>>
> >>>
> >>> --
> >>> -Regards,
> >>> Mayuresh R. Gharat
> >>> (862) 250-7125
> >>> ____________________________________________________________
> >>> What's your flood risk?
> >>> Find flood maps, interactive tools, FAQs, and agents in your area.
> >>>
> http://thirdpartyoffers.netzero.net/TGL3255/55072125266de21244da8mp12duc
> >>
> >> Thanks
> >> Zakee
> >>
> >>
> >>
> >>
> >>
> >
> >
> > --
> > -Regards,
> > Mayuresh R. Gharat
> > (862) 250-7125
> > ____________________________________________________________
> > High School Yearbooks
> > View Class Yearbooks Online Free. Reminisce & Buy a Reprint Today!
> > http://thirdpartyoffers.netzero.net/TGL3255/5507e24f3050f624f0e4amp01duc
>
>


-- 
-Regards,
Mayuresh R. Gharat
(862) 250-7125

Reply via email to