Ditto, filebeat + ELK works very well. You can even tokenize these logs appropriately to have a richer search/filtering.
Cody From: Harsh Choudhary <[email protected]> Reply-To: "[email protected]" <[email protected]> Date: Friday, March 31, 2017 at 4:38 AM To: "[email protected]" <[email protected]> Subject: Re: Centralized logging for storm Hi Shashank What we do is, we have filebeats installed on our Storm clusters and they send the log files data to our central log server, Graylog. This tool is great and you can see your logs like they are one stream of messages, sorted by timestamp. One thing that really helps is that you can also lookup all the other logs near a timestamp. Cheers! Harsh Choudhary On Fri, Mar 31, 2017 at 1:16 PM, Shashank Prasad <[email protected]<mailto:[email protected]>> wrote: Hi folks, Storm is a great tool but the logs are all over the place. As you increase your workers, your log files will increase as well and there is no single file it logs to. This makes it very hard to troubleshoot since you have to tail multiple logs. Ideally, i would like to ship all the logs for a topology to a centralized log server where i could use something like Kibana and filter the logs on what i am searching for. Anyone has any suggestion on how to achieve this or has a use case of how you currently doing it. Thanks a lot for your time! -shashank
