I think log.cleanup.interval.mins was removed in the first 0.8 release. It
sounds like you're looking at outdated docs. Search for
log.retention.check.interval.ms here:
http://kafka.apache.org/documentation.html

As for setting the values too low hurting performance, I'd guess it's
probably only an issue if you set them extremely small, such that file
creation and cleanup become a bottleneck.

-Ewen

On Mon, Jul 27, 2015 at 10:03 AM, Yuheng Du <yuheng.du.h...@gmail.com>
wrote:

> If I want to get higher throughput, should I increase the
> log.segment.bytes?
>
> I don't see log.retention.check.interval.ms, but there is
> log.cleanup.interval.mins, is that what you mean?
>
> If I set log.roll.ms or log.cleanup.interval.mins too small, will it hurt
> the throughput? Thanks.
>
> On Fri, Jul 24, 2015 at 11:03 PM, Ewen Cheslack-Postava <e...@confluent.io
> >
> wrote:
>
> > You'll want to set the log retention policy via
> > log.retention.{ms,minutes,hours} or log.retention.bytes. If you want
> really
> > aggressive collection (e.g., on the order of seconds, as you specified),
> > you might also need to adjust log.segment.bytes/log.roll.{ms,hours} and
> > log.retention.check.interval.ms.
> >
> > On Fri, Jul 24, 2015 at 12:49 PM, Yuheng Du <yuheng.du.h...@gmail.com>
> > wrote:
> >
> > > Hi,
> > >
> > > I am testing the kafka producer performance. So I created a queue and
> > > writes a large amount of data to that queue.
> > >
> > > Is there a way to delete the data automatically after some time, say
> > > whenever the data size reaches 50GB or the retention time exceeds 10
> > > seconds, it will be deleted so my disk won't get filled and new data
> > can't
> > > be written in?
> > >
> > > Thanks.!
> > >
> >
> >
> >
> > --
> > Thanks,
> > Ewen
> >
>



-- 
Thanks,
Ewen

Reply via email to