Evan, >> Is there a setting that controls the log segment size?
http://incubator.apache.org/kafka/configuration.html. Please look at the log.file.size config. >> Better yet, is there a way to tell the broker (maybe from the consumer), look I want you to start a new log segment? No. The broker rolls over to a new log segment when the current one reaches the log.file.size limit. Thanks, Neha On Wed, Dec 7, 2011 at 9:59 AM, Evan Chan <e...@ooyala.com> wrote: > Is there a setting that controls the log segment size? > > Better yet, is there a way to tell the broker (maybe from the consumer), > look I want you to start a new log segment? > > thx, > Evan > > > On Wed, Dec 7, 2011 at 8:09 AM, Neha Narkhede <neha.narkh...@gmail.com > >wrote: > > > Mark, > > > > The SimpleConsumer class has an API for this - getOffsetsBefore(topic, > > partition, time, maxNumOffsets). > > Using this API, you can get the offset "closest" to a particular > timestamp > > at the log segment granularity. By > > log segment granularity, I mean that it doesn't poke inside a log segment > > and find the message right before or > > right after the timestamp you want. > > > > Thus, it is just an approximation. The larger your log segment, the less > > accuracy you will get using this API. > > > > Thanks, > > Neha > > > > On Wed, Dec 7, 2011 at 7:54 AM, Mark <static.void....@gmail.com> wrote: > > > > > Is there any way to consume all messages starting at and/or between two > > > timestamps? > > > > > > > > > -- > -- > *Evan Chan* > Senior Software Engineer | > e...@ooyala.com | (650) 996-4600 > www.ooyala.com | blog <http://www.ooyala.com/blog> | > @ooyala<http://www.twitter.com/ooyala> >