Re: Storm workers get killed in the middle and supervisor restarts
Hi bobby and Martin, Thanks for reach quickly. I am new to storm and do not know how to configure supervisor users. Since martin has filed a jira i will follow if for updates. Thanks again On 7 August 2017 at 20:50, Martin Burian wrote: > Created as https://issues.apache.org/jira/browse/STORM-2682 > M. > > po 7. 8. 2017 v 16:56 odesílatel Bobby Evans napsal: > >> OK Please file JIRA (https://issues.apache.org/jira) under the STORM >> project and I will see what I can do to reproduce the case/fix it. I >> assume that it is not going to be possible to get a snapshot of your >> zookeeper while this is happening? >> >> >> >> - Bobby >> >> >> >> On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian < >> martin.buria...@gmail.com> wrote: >> >> >> The supervisor restarts, but crashes again after some 30 seconds. I tried >> restarting the supervisor and clearing its data, it did not help. >> Martin >> >> po 7. 8. 2017 v 16:34 odesílatel Bobby Evans >> napsal: >> >> No you should not need to configure the users ahead of time. In storm if >> security is turned off the "user" or "owner" of the topology is stored >> internally as the unix user running nimbus. The fact that it is null >> indicates that there is some kind of a bug. I am just trying to gauge how >> serious of a bug it is. Does the supervisor recover after a little while >> or is it stuck in a bad state? >> >> >> >> - Bobby >> >> >> >> On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian < >> martin.buria...@gmail.com> wrote: >> >> >> I am seeing exactly the same exception being thrown by supervisors in my >> topology after an update from storm 1.0.3 to 1.0.4. >> A quick search took me to http://storm.apache.org/ >> releases/1.0.4/SECURITY.html where there are multiple references to some >> "users". Do I need to configure one of those? I Didn't have to do anything >> with users before. >> >> Martin >> >> po 7. 8. 2017 v 16:01 odesílatel Bobby Evans >> napsal: >> >> from the code >> >> https://github.com/apache/storm/blob/v1.1.1/storm-core/ >> src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 >> >> it looks like your user is null when trying to update the resources for >> the topology? Did the supervisor and your topology recover when the >> supervisor was relaunched? If not please file a bug JIRA and we can look >> into it. >> >> >> - Bobby >> >> >> >> On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya < >> sa...@haulmatic.com> wrote: >> >> >> Hi, >> >> We are using apache storm to analyze a GPS data stream by subscribing to >> rabbit mq message channel. We use apache storm 1.11. >> >> We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon >> EC2 instances. Also we have a local storm supervisor which is pointing to >> the remote nimbus in ec2 via the same zookeeper. >> >> >> When we run the topology as a local cluster or submit to the storm when >> only local supervisor is running ( we stop the remote supervisor instance), >> everything works as expected. >> >> Problem arise when we submit the topology to the remote supervisor >> (production like env). This keeps restarting the supervisor with below logs. >> >> >> *worker.log* >> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down >> worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e >> 19f051 6700 >> >> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating >> messaging context >> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down >> executors >> 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting >> down executor fuel-data-analyzer:[2 2] >> 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex >> ecutor[2 2] [INFO] Async loop interrupted! >> >> 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 >> 2]-send-queue [INFO] Async loop interrupted! >> 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down >> executor fuel-data-analyzer:[2 2] >> 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting >> down executor fuel-data-save-to-db:[3 3] >> 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- >> executor[3 3] [INFO] Async loop interrupted! >> >> 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 >> 3]-send-queue [INFO] Async loop interrupted! >> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down >> executor fuel-data-save-to-db:[3 3] >> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting >> down executor __acker:[1 1] >> 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] >> [INFO] Async loop interrupted! >> 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 >> 1]-send-queue [INFO] Async loop interrupted! >> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down >> executor __acker:[1 1] >> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19
Re: Storm workers get killed in the middle and supervisor restarts
Created as https://issues.apache.org/jira/browse/STORM-2682 M. po 7. 8. 2017 v 16:56 odesílatel Bobby Evans napsal: > OK Please file JIRA (https://issues.apache.org/jira) under the STORM > project and I will see what I can do to reproduce the case/fix it. I > assume that it is not going to be possible to get a snapshot of your > zookeeper while this is happening? > > > > - Bobby > > > > On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian < > martin.buria...@gmail.com> wrote: > > > The supervisor restarts, but crashes again after some 30 seconds. I tried > restarting the supervisor and clearing its data, it did not help. > Martin > > po 7. 8. 2017 v 16:34 odesílatel Bobby Evans napsal: > > No you should not need to configure the users ahead of time. In storm if > security is turned off the "user" or "owner" of the topology is stored > internally as the unix user running nimbus. The fact that it is null > indicates that there is some kind of a bug. I am just trying to gauge how > serious of a bug it is. Does the supervisor recover after a little while > or is it stuck in a bad state? > > > > - Bobby > > > > On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian < > martin.buria...@gmail.com> wrote: > > > I am seeing exactly the same exception being thrown by supervisors in my > topology after an update from storm 1.0.3 to 1.0.4. > A quick search took me to > http://storm.apache.org/releases/1.0.4/SECURITY.html where there are > multiple references to some "users". Do I need to configure one of those? I > Didn't have to do anything with users before. > > Martin > > po 7. 8. 2017 v 16:01 odesílatel Bobby Evans napsal: > > from the code > > > https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 > > it looks like your user is null when trying to update the resources for > the topology? Did the supervisor and your topology recover when the > supervisor was relaunched? If not please file a bug JIRA and we can look > into it. > > > - Bobby > > > > On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya < > sa...@haulmatic.com> wrote: > > > Hi, > > We are using apache storm to analyze a GPS data stream by subscribing to > rabbit mq message channel. We use apache storm 1.11. > > We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon > EC2 instances. Also we have a local storm supervisor which is pointing to > the remote nimbus in ec2 via the same zookeeper. > > > When we run the topology as a local cluster or submit to the storm when > only local supervisor is running ( we stop the remote supervisor instance), > everything works as expected. > > Problem arise when we submit the topology to the remote supervisor > (production like env). This keeps restarting the supervisor with below logs. > > > *worker.log* > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e > 19f051 6700 > > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating > messaging context > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > executors > 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex > ecutor[2 2] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 > 2]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- > executor[3 3] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 > 3]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor __acker:[1 1] > 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] > [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 > 1]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down > executor __acker:[1 1] > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor rabbit-mq-gps-reader-spout:[6 6] > 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 > 6]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader > -spout-executor[6 6] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut
Re: Storm workers get killed in the middle and supervisor restarts
OK Please file JIRA (https://issues.apache.org/jira) under the STORM project and I will see what I can do to reproduce the case/fix it. I assume that it is not going to be possible to get a snapshot of your zookeeper while this is happening? - Bobby On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian wrote: The supervisor restarts, but crashes again after some 30 seconds. I tried restarting the supervisor and clearing its data, it did not help.Martin po 7. 8. 2017 v 16:34 odesílatel Bobby Evans napsal: No you should not need to configure the users ahead of time. In storm if security is turned off the "user" or "owner" of the topology is stored internally as the unix user running nimbus. The fact that it is null indicates that there is some kind of a bug. I am just trying to gauge how serious of a bug it is. Does the supervisor recover after a little while or is it stuck in a bad state? - Bobby On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian wrote: I am seeing exactly the same exception being thrown by supervisors in my topology after an update from storm 1.0.3 to 1.0.4.A quick search took me to http://storm.apache.org/releases/1.0.4/SECURITY.html where there are multiple references to some "users". Do I need to configure one of those? I Didn't have to do anything with users before. Martin po 7. 8. 2017 v 16:01 odesílatel Bobby Evans napsal: from the code https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 it looks like your user is null when trying to update the resources for the topology? Did the supervisor and your topology recover when the supervisor was relaunched? If not please file a bug JIRA and we can look into it. - Bobby On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya wrote: Hi, We are using apache storm to analyze a GPS data stream by subscribing to rabbit mq message channel. We use apache storm 1.11. We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 instances. Also we have a local storm supervisor which is pointing to the remote nimbus in ec2 via the same zookeeper. When we run the topology as a local cluster or submit to the storm when only local supervisor is running ( we stop the remote supervisor instance), everything works as expected. Problem arise when we submit the topology to the remote supervisor (production like env). This keeps restarting the supervisor with below logs. worker.log 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 6700 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating messaging context 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down executors 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 2] [INFO] Async loop interrupted! 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 2]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- executor[3 3] [INFO] Async loop interrupted! 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 3]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __acker:[1 1] 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down executor __acker:[1 1] 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 6]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader -spout-executor[6 6] [INFO] Async loop interrupted! 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __system:[-1 -1] 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 -1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] [INFO] Async loop interrupted! 2017
Re: Storm workers get killed in the middle and supervisor restarts
The supervisor restarts, but crashes again after some 30 seconds. I tried restarting the supervisor and clearing its data, it did not help. Martin po 7. 8. 2017 v 16:34 odesílatel Bobby Evans napsal: > No you should not need to configure the users ahead of time. In storm if > security is turned off the "user" or "owner" of the topology is stored > internally as the unix user running nimbus. The fact that it is null > indicates that there is some kind of a bug. I am just trying to gauge how > serious of a bug it is. Does the supervisor recover after a little while > or is it stuck in a bad state? > > > > - Bobby > > > > On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian < > martin.buria...@gmail.com> wrote: > > > I am seeing exactly the same exception being thrown by supervisors in my > topology after an update from storm 1.0.3 to 1.0.4. > A quick search took me to > http://storm.apache.org/releases/1.0.4/SECURITY.html where there are > multiple references to some "users". Do I need to configure one of those? I > Didn't have to do anything with users before. > > Martin > > po 7. 8. 2017 v 16:01 odesílatel Bobby Evans napsal: > > from the code > > > https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 > > it looks like your user is null when trying to update the resources for > the topology? Did the supervisor and your topology recover when the > supervisor was relaunched? If not please file a bug JIRA and we can look > into it. > > > - Bobby > > > > On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya < > sa...@haulmatic.com> wrote: > > > Hi, > > We are using apache storm to analyze a GPS data stream by subscribing to > rabbit mq message channel. We use apache storm 1.11. > > We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon > EC2 instances. Also we have a local storm supervisor which is pointing to > the remote nimbus in ec2 via the same zookeeper. > > > When we run the topology as a local cluster or submit to the storm when > only local supervisor is running ( we stop the remote supervisor instance), > everything works as expected. > > Problem arise when we submit the topology to the remote supervisor > (production like env). This keeps restarting the supervisor with below logs. > > > *worker.log* > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e > 19f051 6700 > > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating > messaging context > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > executors > 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex > ecutor[2 2] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 > 2]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- > executor[3 3] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 > 3]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor __acker:[1 1] > 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] > [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 > 1]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down > executor __acker:[1 1] > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor rabbit-mq-gps-reader-spout:[6 6] > 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 > 6]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader > -spout-executor[6 6] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down > executor rabbit-mq-gps-reader-spout:[6 6] > 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor __system:[-1 -1] > 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 > -1]-send-queue [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] > [INFO] Async loop interrupted! > 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down > executor __system:[-1 -1] > 2017-08
Re: Storm workers get killed in the middle and supervisor restarts
No you should not need to configure the users ahead of time. In storm if security is turned off the "user" or "owner" of the topology is stored internally as the unix user running nimbus. The fact that it is null indicates that there is some kind of a bug. I am just trying to gauge how serious of a bug it is. Does the supervisor recover after a little while or is it stuck in a bad state? - Bobby On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian wrote: I am seeing exactly the same exception being thrown by supervisors in my topology after an update from storm 1.0.3 to 1.0.4.A quick search took me to http://storm.apache.org/releases/1.0.4/SECURITY.html where there are multiple references to some "users". Do I need to configure one of those? I Didn't have to do anything with users before. Martin po 7. 8. 2017 v 16:01 odesílatel Bobby Evans napsal: from the code https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 it looks like your user is null when trying to update the resources for the topology? Did the supervisor and your topology recover when the supervisor was relaunched? If not please file a bug JIRA and we can look into it. - Bobby On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya wrote: Hi, We are using apache storm to analyze a GPS data stream by subscribing to rabbit mq message channel. We use apache storm 1.11. We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 instances. Also we have a local storm supervisor which is pointing to the remote nimbus in ec2 via the same zookeeper. When we run the topology as a local cluster or submit to the storm when only local supervisor is running ( we stop the remote supervisor instance), everything works as expected. Problem arise when we submit the topology to the remote supervisor (production like env). This keeps restarting the supervisor with below logs. worker.log 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 6700 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating messaging context 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down executors 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 2] [INFO] Async loop interrupted! 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 2]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- executor[3 3] [INFO] Async loop interrupted! 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 3]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __acker:[1 1] 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down executor __acker:[1 1] 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 6]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader -spout-executor[6 6] [INFO] Async loop interrupted! 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __system:[-1 -1] 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 -1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down executor __system:[-1 -1] 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down executor gps-data-logger:[5 5] 2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 5]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec utor[5 5] [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down executor gps-data-logger:[5
Re: Storm workers get killed in the middle and supervisor restarts
I am seeing exactly the same exception being thrown by supervisors in my topology after an update from storm 1.0.3 to 1.0.4. A quick search took me to http://storm.apache.org/releases/1.0.4/SECURITY.html where there are multiple references to some "users". Do I need to configure one of those? I Didn't have to do anything with users before. Martin po 7. 8. 2017 v 16:01 odesílatel Bobby Evans napsal: > from the code > > > https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 > > it looks like your user is null when trying to update the resources for > the topology? Did the supervisor and your topology recover when the > supervisor was relaunched? If not please file a bug JIRA and we can look > into it. > > > - Bobby > > > > On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya < > sa...@haulmatic.com> wrote: > > > Hi, > > We are using apache storm to analyze a GPS data stream by subscribing to > rabbit mq message channel. We use apache storm 1.11. > > We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon > EC2 instances. Also we have a local storm supervisor which is pointing to > the remote nimbus in ec2 via the same zookeeper. > > > When we run the topology as a local cluster or submit to the storm when > only local supervisor is running ( we stop the remote supervisor instance), > everything works as expected. > > Problem arise when we submit the topology to the remote supervisor > (production like env). This keeps restarting the supervisor with below logs. > > > *worker.log* > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e > 19f051 6700 > > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating > messaging context > 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down > executors > 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex > ecutor[2 2] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 > 2]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-analyzer:[2 2] > 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- > executor[3 3] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 > 3]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down > executor fuel-data-save-to-db:[3 3] > 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor __acker:[1 1] > 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] > [INFO] Async loop interrupted! > 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 > 1]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down > executor __acker:[1 1] > 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor rabbit-mq-gps-reader-spout:[6 6] > 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 > 6]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader > -spout-executor[6 6] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down > executor rabbit-mq-gps-reader-spout:[6 6] > 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor __system:[-1 -1] > 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 > -1]-send-queue [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] > [INFO] Async loop interrupted! > 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down > executor __system:[-1 -1] > 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor gps-data-logger:[5 5] > 2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 > 5]-send-queue [INFO] Async loop interrupted! > 2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec > utor[5 5] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down > executor gps-data-logger:[5 5] > 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting > down executor gps-data-devider:[4 4] > 2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-exe > cutor[4 4] [INFO] Async loop interrupted! > > 2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4 > 4]-send
Re: Storm workers get killed in the middle and supervisor restarts
from the code https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332 it looks like your user is null when trying to update the resources for the topology? Did the supervisor and your topology recover when the supervisor was relaunched? If not please file a bug JIRA and we can look into it. - Bobby On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya wrote: Hi, We are using apache storm to analyze a GPS data stream by subscribing to rabbit mq message channel. We use apache storm 1.11. We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 instances. Also we have a local storm supervisor which is pointing to the remote nimbus in ec2 via the same zookeeper. When we run the topology as a local cluster or submit to the storm when only local supervisor is running ( we stop the remote supervisor instance), everything works as expected. Problem arise when we submit the topology to the remote supervisor (production like env). This keeps restarting the supervisor with below logs. worker.log 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 6700 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating messaging context 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down executors 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 2] [INFO] Async loop interrupted! 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 2]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- executor[3 3] [INFO] Async loop interrupted! 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 3]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __acker:[1 1] 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down executor __acker:[1 1] 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 6]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader -spout-executor[6 6] [INFO] Async loop interrupted! 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __system:[-1 -1] 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 -1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down executor __system:[-1 -1] 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down executor gps-data-logger:[5 5] 2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 5]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec utor[5 5] [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down executor gps-data-logger:[5 5] 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting down executor gps-data-devider:[4 4] 2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-exe cutor[4 4] [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4 4]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down executor gps-data-devider:[4 4] 2017-08-05 08:50:27.351 o.a.s.d.worker Thread-19 [INFO] Shut down executors 2017-08-05 08:50:27.353 o.a.s.d.worker Thread-19 [INFO] Shutting down transfer thread 2017-08-05 08:50:27.353 o.a.s.util Thread-17-disruptor-worker-tra nsfer-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down transfer thread 2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut dow
Storm workers get killed in the middle and supervisor restarts
Hi, We are using apache storm to analyze a GPS data stream by subscribing to rabbit mq message channel. We use apache storm 1.11. We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 instances. Also we have a local storm supervisor which is pointing to the remote nimbus in ec2 via the same zookeeper. When we run the topology as a local cluster or submit to the storm when only local supervisor is running ( we stop the remote supervisor instance), everything works as expected. Problem arise when we submit the topology to the remote supervisor (production like env). This keeps restarting the supervisor with below logs. *worker.log* 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker GpsDataAnalyticsTopology-6-1501922728 a383ada8-62a4-418c-9c7c-4e5a5e19f051 6700 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating messaging context 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down executors 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-executor[2 2] [INFO] Async loop interrupted! 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 2]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-analyzer:[2 2] 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-executor[3 3] [INFO] Async loop interrupted! 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 3]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down executor fuel-data-save-to-db:[3 3] 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __acker:[1 1] 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down executor __acker:[1 1] 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 6]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader-spout-executor[6 6] [INFO] Async loop interrupted! 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down executor rabbit-mq-gps-reader-spout:[6 6] 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down executor __system:[-1 -1] 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[-1 -1]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down executor __system:[-1 -1] 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down executor gps-data-logger:[5 5] 2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 5]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-executor[5 5] [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down executor gps-data-logger:[5 5] 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting down executor gps-data-devider:[4 4] 2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-executor[4 4] [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4 4]-send-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down executor gps-data-devider:[4 4] 2017-08-05 08:50:27.351 o.a.s.d.worker Thread-19 [INFO] Shut down executors 2017-08-05 08:50:27.353 o.a.s.d.worker Thread-19 [INFO] Shutting down transfer thread 2017-08-05 08:50:27.353 o.a.s.util Thread-17-disruptor-worker-transfer-queue [INFO] Async loop interrupted! 2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down transfer thread 2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down backpressure thread 2017-08-05 08:50:27.355 o.a.s.d.worker Thread-19 [INFO] Shutting down default resources 2017-08-05 08:50:27.356 o.a.s.d.worker Thread-19 [INFO] Shut down default resources 2017-08-05 08:50:27.356 o.a.s.d.worker Thread-19 [INFO] Trigger any worker shutdown hooks 2017-08-05 08:50:27.363 o.a.s.d.worker Thread-19 [INFO] Disconnecting from storm cluster state context 2017-08-05 08:50:27.363 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl Curator-Framework-0 [INFO] backg