Re: how to consume/reset aggregated bucket

2017-03-31 Thread Rajat Gangwar
Let me rephrase my question : Storm keeps on aggregating same bucket after segregating tuples based on grouped fields. - I want to consume all the aggregated content every hour, and reset aggregation count back to zero. Storm should not be updating this bucket while consuming. - Storm then

Re: Centralized logging for storm

2017-03-31 Thread Artem Ervits
If you deploy Storm with Ambari, you can leverage Ambari Log Search, we aggregate and parse the logs and present it in a nice UI. https://docs.hortonworks.com/HDPDocuments/Ambari-2.4.2.0/bk_ambari-user-guide/content/viewing_service_logs.html The other benefits are prebuilt Grafana dashboards,

Re: Centralized logging for storm

2017-03-31 Thread Cody Lee
Also possible, your choice really. We preferred filebeat as it can be configured to be a reliable source for log output (the consequence is during ELK unavailability (network, elk problems, etc) it can be delayed, but you’ll get everything). IIRC the logstash emitter is best effort and will

Re: Centralized logging for storm

2017-03-31 Thread Antoine Tran
How about modifying the worker.xml configuration, so that we add a appender to logstash/elasticsearch ? No need to add a filebeat if this is handled by Storm itself. On 31/03/2017 16:44, Cody Lee wrote: Ditto, filebeat + ELK works very well. You can even tokenize these logs appropriately

Re: Centralized logging for storm

2017-03-31 Thread Cody Lee
Ditto, filebeat + ELK works very well. You can even tokenize these logs appropriately to have a richer search/filtering. Cody From: Harsh Choudhary Reply-To: "user@storm.apache.org" Date: Friday, March 31, 2017 at 4:38 AM To:

Re: Centralized logging for storm

2017-03-31 Thread Harsh Choudhary
Hi Shashank What we do is, we have filebeats installed on our Storm clusters and they send the log files data to our central log server, Graylog. This tool is great and you can see your logs like they are one stream of messages, sorted by timestamp. One thing that really helps is that you can

Re: nimbus.thrift.max_buffer_size

2017-03-31 Thread Xin Wang
Hi Qingxiang, For large topology, you can change the `nimbus.thrift.max_buffer_size` value to 5242880(5MB), default is 1048576(1MB). - Xin 2017-03-31 16:46 GMT+08:00 Antoine Tran : > https://issues.apache.org/jira/browse/STORM-1469 > > On 31/03/2017 05:37, 马庆祥

Re: nimbus.thrift.max_buffer_size

2017-03-31 Thread Antoine Tran
https://issues.apache.org/jira/browse/STORM-1469 On 31/03/2017 05:37, 马庆祥 wrote: Hello all, I have a question about the "nimbus.thrift.max_buffer_size" in storm.yaml. nimbus.thrift.max_buffer_size: 1048576 I want to get the cluster state and topology state by new NimbusClient, but i got a

Re: Centralized logging for storm

2017-03-31 Thread Stefan Rakonjac
Hello Shashank, I understand this won't help with merging all the log outputs into one single file, but you can configure your Storm workers to write logs to an NFS mounted to each of the workers (that's what we do) and than access them all at that one place with whichever tool you prefer.

Re: Centralized logging for storm

2017-03-31 Thread Jungtaek Lim
Hi Shashank, previously I used Fluentd with some plugins to tail log lines, parse, transform, and store to ES. I think you can do it similar with Logstash. There're some frameworks like Apache NiFi to help ETL and I guess it can achieve what you'd like to do. Thanks, Jungtaek Lim (HeartSaVioR)

Centralized logging for storm

2017-03-31 Thread Shashank Prasad
Hi folks, Storm is a great tool but the logs are all over the place. As you increase your workers, your log files will increase as well and there is no single file it logs to. This makes it very hard to troubleshoot since you have to tail multiple logs. Ideally, i would like to ship all the

Re: Storm files and folders permissions on Linux

2017-03-31 Thread Harsh Choudhary
Can you show the exact error you getting? *Cheers!* Harsh Choudhary On Thu, Mar 30, 2017 at 11:39 PM, I PVP wrote: > that is how it is being done as of now: > > sudo chown -R storm:storm /opt/storm > sudo chmod -R 700 /opt/storm > > but still facing some issues while