I also wanted to add that even on my interface it shows:

Your current configuration is 20,000,000 documents per index and a maximum 
number of 20 indices. 

On Wednesday, May 28, 2014 3:41:49 PM UTC-6, Tyler Bell wrote:
>
> Thanks Kay! I did upgrade from an older version of Graylog and 
> ElasticSearch once v0.20.0 came out.
>
> I ran that job from the web interface, and it looks like it cycled through 
> all my indexes with messages like:
>
> 2014-05-28 14:34:30,460 INFO : org.graylog2.system.jobs.SystemJobManager - 
> Submitted SystemJob <da2267c5-e6af-11e3-a691-005056b274fe> 
> [org.graylog2.indexer.ranges.RebuildIndexRangesJob]
> 2014-05-28 14:34:30,460 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Re-calculating index 
> ranges.
> 2014-05-28 14:34:30,729 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Calculated range of 
> [graylog2_56] in [239ms].
> 2014-05-28 14:34:30,959 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Calculated range of 
> [graylog2_55] in [229ms].
> 2014-05-28 14:34:31,158 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Calculated range of 
> [graylog2_54] in [199ms].
> 2014-05-28 14:34:31,345 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Calculated range of 
> [graylog2_53] in [187ms].
>
> .....etc, then when it hits graylog2_recent
>
> 2014-05-28 14:34:51,242 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Could not calculate 
> range of index [graylog2_recent]. Skipping.
> org.elasticsearch.action.search.SearchPhaseExecutionException: Failed to 
> execute phase [query], all shards failed; shardFailures 
> {[cjdL5h8tSo6VK3zE-Gpvxw][graylog2_recent][0]: 
> RemoteTransportException[[GLNode1][inet[/:9300]][search/phase/query]]; 
> nested: SearchParseException[[graylog2_recent][0]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [Failed to parse 
> source 
> [{"size":1,"query":{"match_all":{}},"sort":[{"timestamp":{"order":"desc"}}]}]]];
>  
> nested: SearchParseException[[graylog2_recent][0]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [No mapping found 
> for [timestamp] in order to sort on]]; 
> }{[XB6yIWzpRHOa8wOJtuKe1g][graylog2_recent][3]: 
> RemoteTransportException[[GLNode2][inet[/:9300]][search/phase/query]]; 
> nested: SearchParseException[[graylog2_recent][3]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [Failed to parse 
> source 
> [{"size":1,"query":{"match_all":{}},"sort":[{"timestamp":{"order":"desc"}}]}]]];
>  
> nested: SearchParseException[[graylog2_recent][3]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [No mapping found 
> for [timestamp] in order to sort on]]; 
> }{[XB6yIWzpRHOa8wOJtuKe1g][graylog2_recent][2]: 
> RemoteTransportException[[GLNode2][inet[/:9300]][search/phase/query]]; 
> nested: SearchParseException[[graylog2_recent][2]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [Failed to parse 
> source 
> [{"size":1,"query":{"match_all":{}},"sort":[{"timestamp":{"order":"desc"}}]}]]];
>  
> nested: SearchParseException[[graylog2_recent][2]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [No mapping found 
> for [timestamp] in order to sort on]]; 
> }{[cjdL5h8tSo6VK3zE-Gpvxw][graylog2_recent][1]: 
> RemoteTransportException[[GLNode1][inet[/:9300]][search/phase/query]]; 
> nested: SearchParseException[[graylog2_recent][1]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [Failed to parse 
> source 
> [{"size":1,"query":{"match_all":{}},"sort":[{"timestamp":{"order":"desc"}}]}]]];
>  
> nested: SearchParseException[[graylog2_recent][1]: 
> query[ConstantScore(*:*)],from[-1],size[1]: Parse Failure [No mapping found 
> for [timestamp] in order to sort on]]; }
>         at 
> org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.onFirstPhaseResult(TransportSearchTypeAction.java:272)
>         at 
> org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$3.onFailure(TransportSearchTypeAction.java:224)
>         at 
> org.elasticsearch.search.action.SearchServiceTransportAction$4.handleException(SearchServiceTransportAction.java:222)
>         at 
> org.elasticsearch.transport.netty.MessageChannelHandler.handleException(MessageChannelHandler.java:181)
>         at 
> org.elasticsearch.transport.netty.MessageChannelHandler.handlerResponseError(MessageChannelHandler.java:171)
>         at 
> org.elasticsearch.transport.netty.MessageChannelHandler.messageReceived(MessageChannelHandler.java:123)
>         at 
> org.elasticsearch.common.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>         at 
> org.elasticsearch.common.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:564)
>         at 
> org.elasticsearch.common.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendUpstream(DefaultChannelPipeline.java:791)
>         at 
> org.elasticsearch.common.netty.channel.Channels.fireMessageReceived(Channels.java:296)
>         at 
> org.elasticsearch.common.netty.handler.codec.frame.FrameDecoder.unfoldAndFireMessageReceived(FrameDecoder.java:462)
>         at 
> org.elasticsearch.common.netty.handler.codec.frame.FrameDecoder.callDecode(FrameDecoder.java:443)
>         at 
> org.elasticsearch.common.netty.handler.codec.frame.FrameDecoder.messageReceived(FrameDecoder.java:303)
>         at 
> org.elasticsearch.common.netty.channel.SimpleChannelUpstreamHandler.handleUpstream(SimpleChannelUpstreamHandler.java:70)
>         at 
> org.elasticsearch.common.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:564)
>         at 
> org.elasticsearch.common.netty.channel.DefaultChannelPipeline.sendUpstream(DefaultChannelPipeline.java:559)
>         at 
> org.elasticsearch.common.netty.channel.Channels.fireMessageReceived(Channels.java:268)
>         at 
> org.elasticsearch.common.netty.channel.Channels.fireMessageReceived(Channels.java:255)
>         at 
> org.elasticsearch.common.netty.channel.socket.nio.NioWorker.read(NioWorker.java:88)
>         at 
> org.elasticsearch.common.netty.channel.socket.nio.AbstractNioWorker.process(AbstractNioWorker.java:108)
>         at 
> org.elasticsearch.common.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:318)
>         at 
> org.elasticsearch.common.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:89)
>         at 
> org.elasticsearch.common.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178)
>         at 
> org.elasticsearch.common.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108)
>         at 
> org.elasticsearch.common.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:744)
>
> Then at the end:
>
> 2014-05-28 14:35:05,052 INFO : 
> org.graylog2.indexer.ranges.RebuildIndexRangesJob - Done calculating index 
> ranges for 172 indices. Took 34562ms.
> 2014-05-28 14:35:05,052 INFO : org.graylog2.system.jobs.SystemJobManager - 
> SystemJob <da2267c5-e6af-11e3-a691-005056b274fe> 
> [org.graylog2.indexer.ranges.RebuildIndexRangesJob] finished in 34591ms.
>
> What does that job actually do? Should those older indices be marked for 
> deletion now?
>
>
> On Wednesday, May 28, 2014 3:13:33 PM UTC-6, Kay Röpke wrote:
>>
>> Hi!
>>
>> graylog2_recent sounds like you upgraded from an older version. Those 
>> indices might not get collected correctly.
>>
>> Please try to use the "recalculate index ranges" from the action menu on 
>> the system/indices page in the web interface. After that the server should 
>> perform retention properly. If not, you can still manually clear out 
>> unwanted indices, but be sure to perform that maintenance task from the 
>> page.
>>
>> Best,
>> Kay
>> On May 28, 2014 11:04 PM, "Tyler Bell" <[email protected]> wrote:
>>
>>> Hi All - Using Graylog v0.20.1 and ES 0.90.10, and am having my disk 
>>> space maxed out because indices are not rotating out. My elasticsearch data 
>>> directory shows indices graylog2_recent, graylog2_0, ...., graylog2_173.
>>>
>>> Anyone have experience with this? I'm going to use the ES API to clear 
>>> out some older indices and get my setup working again, but need to figure 
>>> out the rotation issue for longterm resolution.
>>>
>>> I'm using default config settings:
>>>
>>> # Embedded elasticsearch configuration file
>>> # pay attention to the working directory of the server, maybe use an 
>>> absolute path here
>>> elasticsearch_config_file = /etc/graylog2-elasticsearch.yml
>>>
>>> elasticsearch_max_docs_per_index = 20000000
>>>
>>> # How many indices do you want to keep?
>>> # 
>>> elasticsearch_max_number_of_indices*elasticsearch_max_docs_per_index=total 
>>> number of messages in your setup
>>> elasticsearch_max_number_of_indices = 20
>>>
>>> # Decide what happens with the oldest indices when the maximum number of 
>>> indices is reached.
>>> # The following strategies are availble:
>>> #   - delete # Deletes the index completely (Default)
>>> #   - close # Closes the index and hides it from the system. Can be 
>>> re-opened later.
>>> retention_strategy = delete
>>>
>>>  -- 
>>> You received this message because you are subscribed to the Google 
>>> Groups "graylog2" group.
>>> To unsubscribe from this group and stop receiving emails from it, send 
>>> an email to [email protected].
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"graylog2" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to