[jira] [Commented] (KAFKA-3802) log mtimes reset on broker restart

2016-06-07 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-3802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15319321#comment-15319321
 ] 

Andrew Otto commented on KAFKA-3802:


I am running Kafka 0.9.01 on Debian Jessie, with 12 log.dirs on different 
disks.  

Here are my server.logs during the shutdown:

{noformat}
[2016-06-07 17:50:41,421] INFO [Kafka Server 20], shutting down 
(kafka.server.KafkaServer)
[2016-06-07 17:50:41,428] INFO [Kafka Server 20], Starting controlled shutdown 
(kafka.server.KafkaServer)
[2016-06-07 17:50:41,956] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_MobileWebBrowse,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,159] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_MobileWebBrowse,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,161] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [__consumer_offsets,14] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,161] INFO Truncating log __consumer_offsets-14 to offset 
108. (kafka.log.Log)
[2016-06-07 17:50:42,599] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_MobileAppShareAttempts,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,600] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_MobileAppShareAttempts,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,601] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging-valid-mixed,8] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,603] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging-valid-mixed,8] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,603] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_text,9] (kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,605] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_text,9] (kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,606] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [__consumer_offsets,28] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,608] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [__consumer_offsets,28] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,608] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_GuidedTourButtonClick,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,611] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_GuidedTourButtonClick,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,613] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_parsoid,1] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,619] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_parsoid,1] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,621] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_CompletionSuggestions,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,624] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_CompletionSuggestions,0] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,625] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_maps,5] (kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,625] INFO [ReplicaFetcherThread-10-12], Shutting down 
(kafka.server.ReplicaFetcherThread)
[2016-06-07 17:50:42,659] INFO [ReplicaFetcherThread-10-12], Stopped  
(kafka.server.ReplicaFetcherThread)
[2016-06-07 17:50:42,659] INFO [ReplicaFetcherThread-10-12], Shutdown completed 
(kafka.server.ReplicaFetcherThread)
[2016-06-07 17:50:42,663] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [webrequest_maps,5] (kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,667] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging-valid-mixed,3] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,679] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging-valid-mixed,3] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,684] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [__consumer_offsets,36] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,690] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [__consumer_offsets,36] 
(kafka.server.ReplicaFetcherManager)
[2016-06-07 17:50:42,691] INFO [ReplicaFetcherManager on broker 20] Removed 
fetcher for partitions [eventlogging_EventError,0] 

[jira] [Commented] (KAFKA-3802) log mtimes reset on broker restart

2016-06-07 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-3802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15319225#comment-15319225
 ] 

Andrew Otto commented on KAFKA-3802:


Possibly related to KAFKA-1379

> log mtimes reset on broker restart
> --
>
> Key: KAFKA-3802
> URL: https://issues.apache.org/jira/browse/KAFKA-3802
> Project: Kafka
>  Issue Type: Bug
>Affects Versions: 0.9.0.1
>Reporter: Andrew Otto
>
> Folks over in 
> http://mail-archives.apache.org/mod_mbox/kafka-users/201605.mbox/%3CCAO8=cz0ragjad1acx4geqcwj+rkd1gmdavkjwytwthkszfg...@mail.gmail.com%3E
>  are commenting about this issue.
> In 0.9, any data log file that was on
> disk before the broker has it's mtime modified to the time of the broker
> restart.
> This causes problems with log retention, as all the files then look like
> they contain recent data to kafka.  We use the default log retention of 7
> days, but if all the files are touched at the same time, this can cause us
> to retain up to 2 weeks of log data, which can fill up our disks.
> This happens *most* of the time, but seemingly not all.  We have seen broker 
> restarts where mtimes were not changed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (KAFKA-3802) log mtimes reset on broker restart

2016-06-07 Thread Andrew Otto (JIRA)
Andrew Otto created KAFKA-3802:
--

 Summary: log mtimes reset on broker restart
 Key: KAFKA-3802
 URL: https://issues.apache.org/jira/browse/KAFKA-3802
 Project: Kafka
  Issue Type: Bug
Affects Versions: 0.9.0.1
Reporter: Andrew Otto


Folks over in 
http://mail-archives.apache.org/mod_mbox/kafka-users/201605.mbox/%3CCAO8=cz0ragjad1acx4geqcwj+rkd1gmdavkjwytwthkszfg...@mail.gmail.com%3E
 are commenting about this issue.

In 0.9, any data log file that was on
disk before the broker has it's mtime modified to the time of the broker
restart.

This causes problems with log retention, as all the files then look like
they contain recent data to kafka.  We use the default log retention of 7
days, but if all the files are touched at the same time, this can cause us
to retain up to 2 weeks of log data, which can fill up our disks.

This happens *most* of the time, but seemingly not all.  We have seen broker 
restarts where mtimes were not changed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (KAFKA-2143) Replicas get ahead of leader and fail

2015-12-14 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-2143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15056672#comment-15056672
 ] 

Andrew Otto commented on KAFKA-2143:


I'd like to add that Wikimedia is experiencing this as well.  It only started 
happening to us recently when we started producing to Kafka via a PHP process 
serving a web request.  This process is short lived, so only connects to Kafka 
in order to produce a single message, and then disconnects.  The only topic 
that this happens to is the one that this PHP process produces to.

Everything recovers as it should, but something is definitely wrong here.  Why 
is a follower trying to consume an offset ahead of what the leader has?

We're running 0.8.2.1 (with the snappy compression bugfix backported).

> Replicas get ahead of leader and fail
> -
>
> Key: KAFKA-2143
> URL: https://issues.apache.org/jira/browse/KAFKA-2143
> Project: Kafka
>  Issue Type: Bug
>  Components: replication
>Affects Versions: 0.8.2.1
>Reporter: Evan Huus
>Assignee: Jiangjie Qin
>
> On a cluster of 6 nodes, we recently saw a case where a single 
> under-replicated partition suddenly appeared, replication lag spiked, and 
> network IO spiked. The cluster appeared to recover eventually on its own,
> Looking at the logs, the thing which failed was partition 7 of the topic 
> {{background_queue}}. It had an ISR of 1,4,3 and its leader at the time was 
> 3. Here are the interesting log lines:
> On node 3 (the leader):
> {noformat}
> [2015-04-23 16:50:05,879] ERROR [Replica Manager on Broker 3]: Error when 
> processing fetch request for partition [background_queue,7] offset 3722949957 
> from follower with correlation id 148185816. Possible cause: Request for 
> offset 3722949957 but we only have log segments in the range 3648049863 to 
> 3722949955. (kafka.server.ReplicaManager)
> [2015-04-23 16:50:05,879] ERROR [Replica Manager on Broker 3]: Error when 
> processing fetch request for partition [background_queue,7] offset 3722949957 
> from follower with correlation id 156007054. Possible cause: Request for 
> offset 3722949957 but we only have log segments in the range 3648049863 to 
> 3722949955. (kafka.server.ReplicaManager)
> [2015-04-23 16:50:13,960] INFO Partition [background_queue,7] on broker 3: 
> Shrinking ISR for partition [background_queue,7] from 1,4,3 to 3 
> (kafka.cluster.Partition)
> {noformat}
> Note that both replicas suddenly asked for an offset *ahead* of the available 
> offsets.
> And on nodes 1 and 4 (the replicas) many occurrences of the following:
> {noformat}
> [2015-04-23 16:50:05,935] INFO Scheduling log segment 3648049863 for log 
> background_queue-7 for deletion. (kafka.log.Log) (edited)
> {noformat}
> Based on my reading, this looks like the replicas somehow got *ahead* of the 
> leader, asked for an invalid offset, got confused, and re-replicated the 
> entire topic from scratch to recover (this matches our network graphs, which 
> show 3 sending a bunch of data to 1 and 4).
> Taking a stab in the dark at the cause, there appears to be a race condition 
> where replicas can receive a new offset before the leader has committed it 
> and is ready to replicate?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (KAFKA-2189) Snappy compression of message batches less efficient in 0.8.2.1

2015-08-13 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-2189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14695589#comment-14695589
 ] 

Andrew Otto commented on KAFKA-2189:


Hi all,

The Wikimedia Foundation had a serious production issue when we upgraded to 
0.8.2.1 because of this bug.  Snappy compression doesn't work at scale in 
0.8.2.1.  I know 0.8.3 is slated for release soon, maybe you should consider 
doing a 0.8.2.2 release just to get this out there in a stable tag, so that 
others don't run into this issue.


 Snappy compression of message batches less efficient in 0.8.2.1
 ---

 Key: KAFKA-2189
 URL: https://issues.apache.org/jira/browse/KAFKA-2189
 Project: Kafka
  Issue Type: Bug
  Components: build, compression, log
Affects Versions: 0.8.2.1
Reporter: Olson,Andrew
Assignee: Ismael Juma
Priority: Blocker
  Labels: trivial
 Fix For: 0.8.3

 Attachments: KAFKA-2189.patch


 We are using snappy compression and noticed a fairly substantial increase 
 (about 2.25x) in log filesystem space consumption after upgrading a Kafka 
 cluster from 0.8.1.1 to 0.8.2.1. We found that this is caused by messages 
 being seemingly recompressed individually (or possibly with a much smaller 
 buffer or dictionary?) instead of as a batch as sent by producers. We 
 eventually tracked down the change in compression ratio/scope to this [1] 
 commit that updated the snappy version from 1.0.5 to 1.1.1.3. The Kafka 
 client version does not appear to be relevant as we can reproduce this with 
 both the 0.8.1.1 and 0.8.2.1 Producer.
 Here are the log files from our troubleshooting that contain the same set of 
 1000 messages, for batch sizes of 1, 10, 100, and 1000. f9d9b was the last 
 commit with 0.8.1.1-like behavior prior to f5ab8 introducing the issue.
 {noformat}
 -rw-rw-r-- 1 kafka kafka 404967 May 12 11:45 
 /var/kafka2/f9d9b-batch-1-0/.log
 -rw-rw-r-- 1 kafka kafka 119951 May 12 11:45 
 /var/kafka2/f9d9b-batch-10-0/.log
 -rw-rw-r-- 1 kafka kafka  89645 May 12 11:45 
 /var/kafka2/f9d9b-batch-100-0/.log
 -rw-rw-r-- 1 kafka kafka  88279 May 12 11:45 
 /var/kafka2/f9d9b-batch-1000-0/.log
 -rw-rw-r-- 1 kafka kafka 402837 May 12 11:41 
 /var/kafka2/f5ab8-batch-1-0/.log
 -rw-rw-r-- 1 kafka kafka 382437 May 12 11:41 
 /var/kafka2/f5ab8-batch-10-0/.log
 -rw-rw-r-- 1 kafka kafka 364791 May 12 11:41 
 /var/kafka2/f5ab8-batch-100-0/.log
 -rw-rw-r-- 1 kafka kafka 380693 May 12 11:41 
 /var/kafka2/f5ab8-batch-1000-0/.log
 {noformat}
 [1] 
 https://github.com/apache/kafka/commit/f5ab8e1780cf80f267906e3259ad4f9278c32d28
  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (KAFKA-1367) Broker topic metadata not kept in sync with ZooKeeper

2014-09-19 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14141002#comment-14141002
 ] 

Andrew Otto commented on KAFKA-1367:


This happens to me as well.  See:  
https://github.com/edenhill/librdkafka/issues/147


 Broker topic metadata not kept in sync with ZooKeeper
 -

 Key: KAFKA-1367
 URL: https://issues.apache.org/jira/browse/KAFKA-1367
 Project: Kafka
  Issue Type: Bug
Affects Versions: 0.8.0, 0.8.1
Reporter: Ryan Berdeen
  Labels: newbie++
 Attachments: KAFKA-1367.txt


 When a broker is restarted, the topic metadata responses from the brokers 
 will be incorrect (different from ZooKeeper) until a preferred replica leader 
 election.
 In the metadata, it looks like leaders are correctly removed from the ISR 
 when a broker disappears, but followers are not. Then, when a broker 
 reappears, the ISR is never updated.
 I used a variation of the Vagrant setup created by Joe Stein to reproduce 
 this with latest from the 0.8.1 branch: 
 https://github.com/also/kafka/commit/dba36a503a5e22ea039df0f9852560b4fb1e067c



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (KAFKA-1367) Broker topic metadata not kept in sync with ZooKeeper

2014-09-19 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14141017#comment-14141017
 ] 

Andrew Otto commented on KAFKA-1367:


I just updated the librdkafka issue, pasting it here as well:

I noticed that in my case, only 1 of the 4 brokers was ever missing in the ISRs 
reported by Kafka Brokers (via librdkafka). That JIRA indicated that a 
preferred-replica-election should fix the problem. I did this:

controlled-shutdown of offending broker 21. Then actual shutdown of broker 21. 
Once this was done, librdkafka metadata showed the correct ISRs, since this 
offending broker really was not in any ISRs. I then restarted broker 21 and let 
its replicas catch back up. Once it caught up, zookeeper reported that all ISRs 
were in sync. I then checked librdkafka's metadata, and broker 21 was not 
listed in any ISR. I then ran a preferred-replica-election. broker 21 was then 
promoted to leader of some partitions. librdkafka then only showed broker 21 
being in the ISRs for which it was also the leader. Any partition that has a 
replica on broker 21 does not show up in the ISR unless broker 21 is the leader.

  $ kafkacat -L -b analytics1022.eqiad.wmnet  -t webrequest_upload
  Metadata for webrequest_upload (from broker -1: 
analytics1022.eqiad.wmnet:9092/bootstrap):
   4 brokers:
broker 12 at analytics1012.eqiad.wmnet:9092
broker 21 at analytics1021.eqiad.wmnet:9092
broker 22 at analytics1022.eqiad.wmnet:9092
broker 18 at analytics1018.eqiad.wmnet:9092
   1 topics:
topic webrequest_upload with 12 partitions:
  partition 11, leader 12, replicas: 12,21,22, isrs: 12,22
  partition 5, leader 21, replicas: 21,22,12, isrs: 22,12,21
  partition 10, leader 22, replicas: 22,18,21, isrs: 18,22
  partition 7, leader 12, replicas: 12,18,21, isrs: 12,18
  partition 8, leader 18, replicas: 18,22,12, isrs: 12,18,22
  partition 3, leader 12, replicas: 12,22,18, isrs: 12,18,22
  partition 4, leader 18, replicas: 18,21,22, isrs: 18,22
  partition 1, leader 21, replicas: 21,18,22, isrs: 18,22,21
  partition 6, leader 22, replicas: 22,12,18, isrs: 12,18,22
  partition 2, leader 22, replicas: 22,21,12, isrs: 12,22
  partition 9, leader 21, replicas: 21,12,18, isrs: 12,18,21
  partition 0, leader 18, replicas: 18,12,21, isrs: 12,18



vs kafka-topic.sh --describe

  Topic:webrequest_upload   PartitionCount:12   ReplicationFactor:3 
Configs:
Topic: webrequest_uploadPartition: 0Leader: 18  
Replicas: 18,12,21  Isr: 12,18,21
Topic: webrequest_uploadPartition: 1Leader: 21  
Replicas: 21,18,22  Isr: 18,22,21
Topic: webrequest_uploadPartition: 2Leader: 22  
Replicas: 22,21,12  Isr: 12,22,21
Topic: webrequest_uploadPartition: 3Leader: 12  
Replicas: 12,22,18  Isr: 12,18,22
Topic: webrequest_uploadPartition: 4Leader: 18  
Replicas: 18,21,22  Isr: 18,22,21
Topic: webrequest_uploadPartition: 5Leader: 21  
Replicas: 21,22,12  Isr: 22,12,21
Topic: webrequest_uploadPartition: 6Leader: 22  
Replicas: 22,12,18  Isr: 12,18,22
Topic: webrequest_uploadPartition: 7Leader: 12  
Replicas: 12,18,21  Isr: 12,18,21
Topic: webrequest_uploadPartition: 8Leader: 18  
Replicas: 18,22,12  Isr: 12,18,22
Topic: webrequest_uploadPartition: 9Leader: 21  
Replicas: 21,12,18  Isr: 12,18,21
Topic: webrequest_uploadPartition: 10   Leader: 22  
Replicas: 22,18,21  Isr: 18,22,21
Topic: webrequest_uploadPartition: 11   Leader: 12  
Replicas: 12,21,22  Isr: 12,22,21

 Broker topic metadata not kept in sync with ZooKeeper
 -

 Key: KAFKA-1367
 URL: https://issues.apache.org/jira/browse/KAFKA-1367
 Project: Kafka
  Issue Type: Bug
Affects Versions: 0.8.0, 0.8.1
Reporter: Ryan Berdeen
  Labels: newbie++
 Attachments: KAFKA-1367.txt


 When a broker is restarted, the topic metadata responses from the brokers 
 will be incorrect (different from ZooKeeper) until a preferred replica leader 
 election.
 In the metadata, it looks like leaders are correctly removed from the ISR 
 when a broker disappears, but followers are not. Then, when a broker 
 reappears, the ISR is never updated.
 I used a variation of the Vagrant setup created by Joe Stein to reproduce 
 this with latest from the 0.8.1 branch: 
 https://github.com/also/kafka/commit/dba36a503a5e22ea039df0f9852560b4fb1e067c



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (KAFKA-1367) Broker topic metadata not kept in sync with ZooKeeper

2014-09-19 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14141017#comment-14141017
 ] 

Andrew Otto edited comment on KAFKA-1367 at 9/19/14 6:15 PM:
-

I just updated the librdkafka issue, pasting it here as well:

I noticed that in my case, only 1 of the 4 brokers was ever missing in the ISRs 
reported by Kafka Brokers (via librdkafka). This JIRA indicated that a 
preferred-replica-election should fix the problem. I did this:

controlled-shutdown of offending broker 21. Then actual shutdown of broker 21. 
Once this was done, librdkafka metadata showed the correct ISRs, since this 
offending broker really was not in any ISRs. I then restarted broker 21 and let 
its replicas catch back up. Once it caught up, zookeeper reported that all ISRs 
were in sync. I then checked librdkafka's metadata, and broker 21 was not 
listed in any ISR. I then ran a preferred-replica-election. broker 21 was then 
promoted to leader of some partitions. librdkafka then only showed broker 21 
being in the ISRs for which it was also the leader. Any partition that has a 
replica on broker 21 does not show up in the ISR unless broker 21 is the leader.

  $ kafkacat -L -b analytics1022.eqiad.wmnet  -t webrequest_upload
  Metadata for webrequest_upload (from broker -1: 
analytics1022.eqiad.wmnet:9092/bootstrap):
   4 brokers:
broker 12 at analytics1012.eqiad.wmnet:9092
broker 21 at analytics1021.eqiad.wmnet:9092
broker 22 at analytics1022.eqiad.wmnet:9092
broker 18 at analytics1018.eqiad.wmnet:9092
   1 topics:
topic webrequest_upload with 12 partitions:
  partition 11, leader 12, replicas: 12,21,22, isrs: 12,22
  partition 5, leader 21, replicas: 21,22,12, isrs: 22,12,21
  partition 10, leader 22, replicas: 22,18,21, isrs: 18,22
  partition 7, leader 12, replicas: 12,18,21, isrs: 12,18
  partition 8, leader 18, replicas: 18,22,12, isrs: 12,18,22
  partition 3, leader 12, replicas: 12,22,18, isrs: 12,18,22
  partition 4, leader 18, replicas: 18,21,22, isrs: 18,22
  partition 1, leader 21, replicas: 21,18,22, isrs: 18,22,21
  partition 6, leader 22, replicas: 22,12,18, isrs: 12,18,22
  partition 2, leader 22, replicas: 22,21,12, isrs: 12,22
  partition 9, leader 21, replicas: 21,12,18, isrs: 12,18,21
  partition 0, leader 18, replicas: 18,12,21, isrs: 12,18



vs kafka-topic.sh --describe

  Topic:webrequest_upload   PartitionCount:12   ReplicationFactor:3 
Configs:
Topic: webrequest_uploadPartition: 0Leader: 18  
Replicas: 18,12,21  Isr: 12,18,21
Topic: webrequest_uploadPartition: 1Leader: 21  
Replicas: 21,18,22  Isr: 18,22,21
Topic: webrequest_uploadPartition: 2Leader: 22  
Replicas: 22,21,12  Isr: 12,22,21
Topic: webrequest_uploadPartition: 3Leader: 12  
Replicas: 12,22,18  Isr: 12,18,22
Topic: webrequest_uploadPartition: 4Leader: 18  
Replicas: 18,21,22  Isr: 18,22,21
Topic: webrequest_uploadPartition: 5Leader: 21  
Replicas: 21,22,12  Isr: 22,12,21
Topic: webrequest_uploadPartition: 6Leader: 22  
Replicas: 22,12,18  Isr: 12,18,22
Topic: webrequest_uploadPartition: 7Leader: 12  
Replicas: 12,18,21  Isr: 12,18,21
Topic: webrequest_uploadPartition: 8Leader: 18  
Replicas: 18,22,12  Isr: 12,18,22
Topic: webrequest_uploadPartition: 9Leader: 21  
Replicas: 21,12,18  Isr: 12,18,21
Topic: webrequest_uploadPartition: 10   Leader: 22  
Replicas: 22,18,21  Isr: 18,22,21
Topic: webrequest_uploadPartition: 11   Leader: 12  
Replicas: 12,21,22  Isr: 12,22,21


was (Author: ottomata):
I just updated the librdkafka issue, pasting it here as well:

I noticed that in my case, only 1 of the 4 brokers was ever missing in the ISRs 
reported by Kafka Brokers (via librdkafka). That JIRA indicated that a 
preferred-replica-election should fix the problem. I did this:

controlled-shutdown of offending broker 21. Then actual shutdown of broker 21. 
Once this was done, librdkafka metadata showed the correct ISRs, since this 
offending broker really was not in any ISRs. I then restarted broker 21 and let 
its replicas catch back up. Once it caught up, zookeeper reported that all ISRs 
were in sync. I then checked librdkafka's metadata, and broker 21 was not 
listed in any ISR. I then ran a preferred-replica-election. broker 21 was then 
promoted to leader of some partitions. librdkafka then only showed broker 21 
being in the ISRs for which it was also the leader. Any partition that has a 
replica on broker 21 does not show up in the ISR unless broker 21 is the leader.

  $ kafkacat -L -b 

[jira] [Created] (KAFKA-1564) Replace barrage of bin/*.sh with single cli wrapper

2014-07-31 Thread Andrew Otto (JIRA)
Andrew Otto created KAFKA-1564:
--

 Summary: Replace barrage of bin/*.sh with single cli wrapper
 Key: KAFKA-1564
 URL: https://issues.apache.org/jira/browse/KAFKA-1564
 Project: Kafka
  Issue Type: Improvement
  Components: packaging
Affects Versions: 0.8.1.1
Reporter: Andrew Otto
Priority: Trivial


Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
small shell wrappers around Java CLIs.  These work, but they are inconvenient 
to package, and not particularly elegant.

If there is interest, I'd like to upstream this script:
https://github.com/wikimedia/operations-debs-kafka/blob/debian/debian/bin/kafka

Wikimedia and Yelp are both using it in production, instead of the provided 
bin/*.sh scripts.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (KAFKA-1564) Replace barrage of bin/*.sh with single cli wrapper

2014-07-31 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14081253#comment-14081253
 ] 

Andrew Otto commented on KAFKA-1564:


Ha, actually, don't quote me on the Yelp is using this in production, all I 
know is that they are submitting patches to this script! :)

 Replace barrage of bin/*.sh with single cli wrapper
 ---

 Key: KAFKA-1564
 URL: https://issues.apache.org/jira/browse/KAFKA-1564
 Project: Kafka
  Issue Type: Improvement
  Components: packaging
Affects Versions: 0.8.1.1
Reporter: Andrew Otto
Priority: Trivial
   Original Estimate: 24h
  Remaining Estimate: 24h

 Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
 small shell wrappers around Java CLIs.  These work, but they are inconvenient 
 to package, and not particularly elegant.
 If there is interest, I'd like to upstream this script:
 https://github.com/wikimedia/operations-debs-kafka/blob/debian/debian/bin/kafka
 Wikimedia and Yelp are both using it in production, instead of the provided 
 bin/*.sh scripts.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (KAFKA-1564) Replace barrage of bin/*.sh with single cli wrapper

2014-07-31 Thread Andrew Otto (JIRA)

 [ 
https://issues.apache.org/jira/browse/KAFKA-1564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Otto updated KAFKA-1564:
---

Description: 
Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
small shell wrappers around Java CLIs.  These work, but they are inconvenient 
to package, and not particularly elegant.

If there is interest, I'd like to upstream this script:
https://github.com/wikimedia/operations-debs-kafka/blob/master/debian/bin/kafka

Wikimedia and Yelp are both using it in production, instead of the provided 
bin/*.sh scripts.

  was:
Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
small shell wrappers around Java CLIs.  These work, but they are inconvenient 
to package, and not particularly elegant.

If there is interest, I'd like to upstream this script:
https://github.com/wikimedia/operations-debs-kafka/blob/debian/debian/bin/kafka

Wikimedia and Yelp are both using it in production, instead of the provided 
bin/*.sh scripts.


 Replace barrage of bin/*.sh with single cli wrapper
 ---

 Key: KAFKA-1564
 URL: https://issues.apache.org/jira/browse/KAFKA-1564
 Project: Kafka
  Issue Type: Improvement
  Components: packaging
Affects Versions: 0.8.1.1
Reporter: Andrew Otto
Priority: Trivial
   Original Estimate: 24h
  Remaining Estimate: 24h

 Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
 small shell wrappers around Java CLIs.  These work, but they are inconvenient 
 to package, and not particularly elegant.
 If there is interest, I'd like to upstream this script:
 https://github.com/wikimedia/operations-debs-kafka/blob/master/debian/bin/kafka
 Wikimedia and Yelp are both using it in production, instead of the provided 
 bin/*.sh scripts.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (KAFKA-1564) Replace barrage of bin/*.sh with single cli wrapper

2014-07-31 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14081259#comment-14081259
 ] 

Andrew Otto commented on KAFKA-1564:


The difficult part might be making the CLASSPATH generic.  Currently it is 
hardcoded for *.jars provided by our .deb.

 Replace barrage of bin/*.sh with single cli wrapper
 ---

 Key: KAFKA-1564
 URL: https://issues.apache.org/jira/browse/KAFKA-1564
 Project: Kafka
  Issue Type: Improvement
  Components: packaging
Affects Versions: 0.8.1.1
Reporter: Andrew Otto
Priority: Trivial
   Original Estimate: 24h
  Remaining Estimate: 24h

 Kafka ships with a bunch of shell wrapper scripts in bin/.  These are very 
 small shell wrappers around Java CLIs.  These work, but they are inconvenient 
 to package, and not particularly elegant.
 If there is interest, I'd like to upstream this script:
 https://github.com/wikimedia/operations-debs-kafka/blob/master/debian/bin/kafka
 Wikimedia and Yelp are both using it in production, instead of the provided 
 bin/*.sh scripts.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (KAFKA-1123) Broker IPv6 addresses parsed incorrectly

2013-11-06 Thread Andrew Otto (JIRA)
Andrew Otto created KAFKA-1123:
--

 Summary: Broker IPv6 addresses parsed incorrectly
 Key: KAFKA-1123
 URL: https://issues.apache.org/jira/browse/KAFKA-1123
 Project: Kafka
  Issue Type: Bug
  Components: producer 
Affects Versions: 0.8
Reporter: Andrew Otto
Assignee: Jun Rao
Priority: Minor


It seems that broker addresses are parsed incorrectly when IPv6 addresses are 
supplied.  IPv6 addresses have colons in them, and Kafka seems to be 
interpreting the first : as the address:port separator.

I have only tried this with the console-producer --broker-list option, so I 
don't know if this affects anything deeper than the CLI.




--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (KAFKA-1046) Added support for Scala 2.10 builds while maintaining compatibility with 2.8.x

2013-09-17 Thread Andrew Otto (JIRA)

[ 
https://issues.apache.org/jira/browse/KAFKA-1046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13769853#comment-13769853
 ] 

Andrew Otto commented on KAFKA-1046:


Hm, I seem to be getting the same compilation error that is in the attached 
Screen Shot.  How do I fix?

 Added support for Scala 2.10 builds while maintaining compatibility with 2.8.x
 --

 Key: KAFKA-1046
 URL: https://issues.apache.org/jira/browse/KAFKA-1046
 Project: Kafka
  Issue Type: Improvement
Affects Versions: 0.8
Reporter: Christopher Freeman
Assignee: Christopher Freeman
 Fix For: 0.8

 Attachments: kafka_2_10_refactor_0.8.patch, 
 kafka_2_10_refactor.patch, Screen Shot 2013-09-09 at 9.34.09 AM.png


 I refactored the project such that it will compile against Scala 2.10.1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira