I'm experiencing a similar issue to this. We have two clusters:
- 2 node monitoring cluster (1 master/data & 1 just data)
- 5 node production cluster (2 data, 3 masters)
The output below is from the non-master data node of the Marvel monitoring
cluster. There are no errors being reported by any of the production nodes.
[2014-08-26 21:10:51,503][DEBUG][action.search.type ]
[stage-search-marvel-1c] [.marvel-2014.08.26][2],
node[iGRH8Gc2QO698RMlWy8rgQ], [P], s[STARTED]: Failed to execute
[org.elasticsearch.action.search.SearchRequest@355e93ff]
org.elasticsearch.transport.RemoteTransportException:
[stage-search-marvel-1b][inet[/10.99.111.122:9300]][search/phase/query]
Caused by: org.elasticsearch.search.SearchParseException:
[.marvel-2014.08.26][2]: query[ConstantScore(BooleanFilter(+*:*
+cache(_type:index_stats) +cache(@timestamp:[1409086800000 TO
1409087460000])))],from[-1],size[10]: Parse Failure [Failed to parse source
[{"size":10,"query":{"filtered":{"query":{"match_all":{}},"filter":{"bool":{"must":[{"match_all":{}},{"term":{"_type":"index_stats"}},{"range":{"@timestamp":{"from":"now-10m/m","to":"now/m"}}}]}}}},"facets":{"timestamp":{"terms_stats":{"key_field":"index.raw","value_field":"@timestamp","order":"term","size":2000}},"primaries.docs.count":{"terms_stats":{"key_field":"index.raw","value_field":"primaries.docs.count","order":"term","size":2000}},"primaries.indexing.index_total":{"terms_stats":{"key_field":"index.raw","value_field":"primaries.indexing.index_total","order":"term","size":2000}},"total.search.query_total":{"terms_stats":{"key_field":"index.raw","value_field":"total.search.query_total","order":"term","size":2000}},"total.merges.total_size_in_bytes":{"terms_stats":{"key_field":"index.raw","value_field":"total.merges.total_size_in_bytes","order":"term","size":2000}},"total.fielddata.memory_size_in_bytes":{"terms_stats":{"key_field":"index.raw","value_field":"total.fielddata.memory_size_in_bytes","order":"term","size":2000}}}}]]
at
org.elasticsearch.search.SearchService.parseSource(SearchService.java:664)
at
org.elasticsearch.search.SearchService.createContext(SearchService.java:515)
at
org.elasticsearch.search.SearchService.createAndPutContext(SearchService.java:487)
at
org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:256)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:688)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:677)
at
org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:275)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.elasticsearch.search.facet.FacetPhaseExecutionException:
Facet [timestamp]: failed to find mapping for index.raw
at
org.elasticsearch.search.facet.termsstats.TermsStatsFacetParser.parse(TermsStatsFacetParser.java:126)
at
org.elasticsearch.search.facet.FacetParseElement.parse(FacetParseElement.java:93)
at
org.elasticsearch.search.SearchService.parseSource(SearchService.java:648)
... 9 more
[2014-08-26 21:10:51,503][DEBUG][action.search.type ]
[stage-search-marvel-1c] [.marvel-2014.08.26][2],
node[iGRH8Gc2QO698RMlWy8rgQ], [P], s[STARTED]: Failed to execute
[org.elasticsearch.action.search.SearchRequest@32f235e9]
org.elasticsearch.transport.RemoteTransportException:
[stage-search-marvel-1b][inet[/10.99.111.122:9300]][search/phase/query]
Caused by: org.elasticsearch.search.SearchParseException:
[.marvel-2014.08.26][2]: query[ConstantScore(BooleanFilter(+*:*
+cache(_type:node_stats) +cache(@timestamp:[1409086800000 TO
1409087460000])))],from[-1],size[10]: Parse Failure [Failed to parse source
[{"size":10,"query":{"filtered":{"query":{"match_all":{}},"filter":{"bool":{"must":[{"match_all":{}},{"term":{"_type":"node_stats"}},{"range":{"@timestamp":{"from":"now-10m/m","to":"now/m"}}}]}}}},"facets":{"timestamp":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"@timestamp","order":"term","size":2000}},"master_nodes":{"terms":{"field":"node.ip_port.raw","size":2000},"facet_filter":{"term":{"node.master":"true"}}},"os.cpu.usage":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"os.cpu.usage","order":"term","size":2000}},"os.load_average.1m":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"os.load_average.1m","order":"term","size":2000}},"jvm.mem.heap_used_percent":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"jvm.mem.heap_used_percent","order":"term","size":2000}},"fs.total.available_in_bytes":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"fs.total.available_in_bytes","order":"term","size":2000}},"fs.total.disk_io_op":{"terms_stats":{"key_field":"node.ip_port.raw","value_field":"fs.total.disk_io_op","order":"term","size":2000}}}}]]
at
org.elasticsearch.search.SearchService.parseSource(SearchService.java:664)
at
org.elasticsearch.search.SearchService.createContext(SearchService.java:515)
at
org.elasticsearch.search.SearchService.createAndPutContext(SearchService.java:487)
at
org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:256)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:688)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:677)
at
org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:275)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.elasticsearch.search.facet.FacetPhaseExecutionException:
Facet [timestamp]: failed to find mapping for node.ip_port.raw
at
org.elasticsearch.search.facet.termsstats.TermsStatsFacetParser.parse(TermsStatsFacetParser.java:126)
at
org.elasticsearch.search.facet.FacetParseElement.parse(FacetParseElement.java:93)
at
org.elasticsearch.search.SearchService.parseSource(SearchService.java:648)
... 9 more
[2014-08-26 21:10:51,504][DEBUG][action.search.type ]
[stage-search-marvel-1c] All shards failed for phase: [query]
On Wednesday, May 28, 2014 10:57:18 AM UTC-4, Prometheus WillSurvive wrote:
>
> Hi,
>
> We have development test cluster 8 nodes running on 0.90.10 ES. Also we
> have one monitoring node separated from dev cluster for marvel.
>
> We installed all nodes to marvel latest /1.1.1/ pointing to monitoring
> marvel IP in the ES config. Told the monitoring node not collect its data.
>
> We restarted all nodes.
>
> Waited hours......
>
> we have only :
>
> *CLUSTER SUMMARY*
>
> - *Name:* indictoYeni*Status:* green*Nodes:* 8*Indices:* 15*Shards:* 75
> *Data:* 1.47 TB*CPU:* 1311%*Memory:* 111.68 GB / 235.20 GB*Up time:*
> 19.9 h*Version:* 0.90.10
>
> But All other information is not available. It says in the dashboard below
> panel :
>
> *Oops!* FacetPhaseExecutionException[Facet [timestamp]: failed to find
> mapping for node.ip_port.raw]
>
>
> and
>
>
> *Oops!* FacetPhaseExecutionException[Facet [timestamp]: failed to find
> mapping for index.raw]
>
>
> I couldnt find a fix.
>
>
> Is there any solution to this ?
>
>
> Thanks
>
> PS
>
>
>
>
>
>
--
You received this message because you are subscribed to the Google Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/elasticsearch/2940771a-7fbd-45f2-a617-3fcea36fff14%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.