Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Sahan Maldeniya
Hi bobby and Martin,

Thanks for reach quickly. I am new to storm and do not know how to
configure supervisor users. Since martin has filed a jira i will follow if
for updates.

Thanks again

On 7 August 2017 at 20:50, Martin Burian  wrote:

> Created as https://issues.apache.org/jira/browse/STORM-2682
> M.
>
> po 7. 8. 2017 v 16:56 odesílatel Bobby Evans  napsal:
>
>> OK Please file JIRA (https://issues.apache.org/jira) under the STORM
>> project and I will see what I can do to reproduce the case/fix it.  I
>> assume that it is not going to be possible to get a snapshot of your
>> zookeeper while this is happening?
>>
>>
>>
>> - Bobby
>>
>>
>>
>> On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian <
>> martin.buria...@gmail.com> wrote:
>>
>>
>> The supervisor restarts, but crashes again after some 30 seconds. I tried
>> restarting the supervisor and clearing its data, it did not help.
>> Martin
>>
>> po 7. 8. 2017 v 16:34 odesílatel Bobby Evans 
>> napsal:
>>
>> No you should not need to configure the users ahead of time.  In storm if
>> security is turned off the "user" or "owner" of the topology is stored
>> internally as the unix user running nimbus.  The fact that it is null
>> indicates that there is some kind of a bug.  I am just trying to gauge how
>> serious of a bug it is.  Does the supervisor recover after a little while
>> or is it stuck in a bad state?
>>
>>
>>
>> - Bobby
>>
>>
>>
>> On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian <
>> martin.buria...@gmail.com> wrote:
>>
>>
>> I am seeing exactly the same exception being thrown by supervisors in my
>> topology after an update from storm 1.0.3 to 1.0.4.
>> A quick search took me to http://storm.apache.org/
>> releases/1.0.4/SECURITY.html where there are multiple references to some
>> "users". Do I need to configure one of those? I Didn't have to do anything
>> with users before.
>>
>> Martin
>>
>> po 7. 8. 2017 v 16:01 odesílatel Bobby Evans 
>> napsal:
>>
>> from the code
>>
>> https://github.com/apache/storm/blob/v1.1.1/storm-core/
>> src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
>>
>> it looks like your user is null when trying to update the resources for
>> the topology?  Did the supervisor and your topology recover when the
>> supervisor was relaunched?  If not please file a bug JIRA and we can look
>> into it.
>>
>>
>> - Bobby
>>
>>
>>
>> On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya <
>> sa...@haulmatic.com> wrote:
>>
>>
>> Hi,
>>
>> We are using apache storm to analyze a GPS data stream by subscribing to
>> rabbit mq message channel. We use apache storm 1.11.
>>
>> We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon
>> EC2 instances. Also we have a local storm supervisor which is pointing to
>> the remote nimbus in ec2 via the same zookeeper.
>>
>>
>> When we run the topology as a local cluster or submit to the storm when
>> only local supervisor is running ( we stop the remote supervisor instance),
>> everything works as expected.
>>
>> Problem arise when we submit the topology to the remote supervisor
>> (production like env). This keeps restarting the supervisor with below logs.
>>
>>
>> *worker.log*
>> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
>> worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e
>> 19f051 6700
>>
>> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating
>> messaging context
>> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
>> executors
>> 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting
>> down executor fuel-data-analyzer:[2 2]
>> 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex
>> ecutor[2 2] [INFO] Async loop interrupted!
>>
>> 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2
>> 2]-send-queue [INFO] Async loop interrupted!
>> 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down
>> executor fuel-data-analyzer:[2 2]
>> 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting
>> down executor fuel-data-save-to-db:[3 3]
>> 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-
>> executor[3 3] [INFO] Async loop interrupted!
>>
>> 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3
>> 3]-send-queue [INFO] Async loop interrupted!
>> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down
>> executor fuel-data-save-to-db:[3 3]
>> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting
>> down executor __acker:[1 1]
>> 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1]
>> [INFO] Async loop interrupted!
>> 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1
>> 1]-send-queue [INFO] Async loop interrupted!
>> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down
>> executor __acker:[1 1]
>> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Martin Burian
Created as https://issues.apache.org/jira/browse/STORM-2682
M.

po 7. 8. 2017 v 16:56 odesílatel Bobby Evans  napsal:

> OK Please file JIRA (https://issues.apache.org/jira) under the STORM
> project and I will see what I can do to reproduce the case/fix it.  I
> assume that it is not going to be possible to get a snapshot of your
> zookeeper while this is happening?
>
>
>
> - Bobby
>
>
>
> On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian <
> martin.buria...@gmail.com> wrote:
>
>
> The supervisor restarts, but crashes again after some 30 seconds. I tried
> restarting the supervisor and clearing its data, it did not help.
> Martin
>
> po 7. 8. 2017 v 16:34 odesílatel Bobby Evans  napsal:
>
> No you should not need to configure the users ahead of time.  In storm if
> security is turned off the "user" or "owner" of the topology is stored
> internally as the unix user running nimbus.  The fact that it is null
> indicates that there is some kind of a bug.  I am just trying to gauge how
> serious of a bug it is.  Does the supervisor recover after a little while
> or is it stuck in a bad state?
>
>
>
> - Bobby
>
>
>
> On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian <
> martin.buria...@gmail.com> wrote:
>
>
> I am seeing exactly the same exception being thrown by supervisors in my
> topology after an update from storm 1.0.3 to 1.0.4.
> A quick search took me to
> http://storm.apache.org/releases/1.0.4/SECURITY.html where there are
> multiple references to some "users". Do I need to configure one of those? I
> Didn't have to do anything with users before.
>
> Martin
>
> po 7. 8. 2017 v 16:01 odesílatel Bobby Evans  napsal:
>
> from the code
>
>
> https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
>
> it looks like your user is null when trying to update the resources for
> the topology?  Did the supervisor and your topology recover when the
> supervisor was relaunched?  If not please file a bug JIRA and we can look
> into it.
>
>
> - Bobby
>
>
>
> On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya <
> sa...@haulmatic.com> wrote:
>
>
> Hi,
>
> We are using apache storm to analyze a GPS data stream by subscribing to
> rabbit mq message channel. We use apache storm 1.11.
>
> We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon
> EC2 instances. Also we have a local storm supervisor which is pointing to
> the remote nimbus in ec2 via the same zookeeper.
>
>
> When we run the topology as a local cluster or submit to the storm when
> only local supervisor is running ( we stop the remote supervisor instance),
> everything works as expected.
>
> Problem arise when we submit the topology to the remote supervisor
> (production like env). This keeps restarting the supervisor with below logs.
>
>
> *worker.log*
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e
> 19f051 6700
>
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating
> messaging context
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> executors
> 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex
> ecutor[2 2] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2
> 2]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-
> executor[3 3] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3
> 3]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor __acker:[1 1]
> 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1]
> [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1
> 1]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor __acker:[1 1]
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor rabbit-mq-gps-reader-spout:[6 6]
> 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6
> 6]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader
> -spout-executor[6 6] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut 

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Bobby Evans
OK Please file JIRA (https://issues.apache.org/jira) under the STORM project 
and I will see what I can do to reproduce the case/fix it.  I assume that it is 
not going to be possible to get a snapshot of your zookeeper while this is 
happening?


- Bobby


On Monday, August 7, 2017, 9:53:34 AM CDT, Martin Burian 
 wrote:

The supervisor restarts, but crashes again after some 30 seconds. I tried 
restarting the supervisor and clearing its data, it did not help.Martin
po 7. 8. 2017 v 16:34 odesílatel Bobby Evans  napsal:

No you should not need to configure the users ahead of time.  In storm if 
security is turned off the "user" or "owner" of the topology is stored 
internally as the unix user running nimbus.  The fact that it is null indicates 
that there is some kind of a bug.  I am just trying to gauge how serious of a 
bug it is.  Does the supervisor recover after a little while or is it stuck in 
a bad state?


- Bobby


On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian 
 wrote:

I am seeing exactly the same exception being thrown by supervisors in my 
topology after an update from storm 1.0.3 to 1.0.4.A quick search took me to 
http://storm.apache.org/releases/1.0.4/SECURITY.html where there are multiple 
references to some "users". Do I need to configure one of those? I Didn't have 
to do anything with users before.
Martin

po 7. 8. 2017 v 16:01 odesílatel Bobby Evans  napsal:

from the code
https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
it looks like your user is null when trying to update the resources for the 
topology?  Did the supervisor and your topology recover when the supervisor was 
relaunched?  If not please file a bug JIRA and we can look into it.


- Bobby


On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya 
 wrote:


Hi,

We are using apache storm to analyze a GPS data stream by subscribing to rabbit 
mq message channel. We use apache storm 1.11.

We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 
instances. Also we have a local storm supervisor which is pointing to the 
remote nimbus in ec2 via the same zookeeper. 





When we run the topology as a local cluster or submit to the storm when only 
local supervisor is running ( we stop the remote supervisor instance), 
everything works as expected.

Problem arise when we submit the topology to the remote supervisor (production 
like env). This keeps restarting the supervisor with below logs.




worker.log
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker 
GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 
6700
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating 
messaging context
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down 
executors
    2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 
2] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 
2]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- 
executor[3 3] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 
3]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __acker:[1 1]
    2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] 
Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 
1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor __acker:[1 1]
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 
6]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader 
-spout-executor[6 6] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __system:[-1 -1]
    2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 
-1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] 
[INFO] Async loop interrupted!
    2017

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Martin Burian
The supervisor restarts, but crashes again after some 30 seconds. I tried
restarting the supervisor and clearing its data, it did not help.
Martin

po 7. 8. 2017 v 16:34 odesílatel Bobby Evans  napsal:

> No you should not need to configure the users ahead of time.  In storm if
> security is turned off the "user" or "owner" of the topology is stored
> internally as the unix user running nimbus.  The fact that it is null
> indicates that there is some kind of a bug.  I am just trying to gauge how
> serious of a bug it is.  Does the supervisor recover after a little while
> or is it stuck in a bad state?
>
>
>
> - Bobby
>
>
>
> On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian <
> martin.buria...@gmail.com> wrote:
>
>
> I am seeing exactly the same exception being thrown by supervisors in my
> topology after an update from storm 1.0.3 to 1.0.4.
> A quick search took me to
> http://storm.apache.org/releases/1.0.4/SECURITY.html where there are
> multiple references to some "users". Do I need to configure one of those? I
> Didn't have to do anything with users before.
>
> Martin
>
> po 7. 8. 2017 v 16:01 odesílatel Bobby Evans  napsal:
>
> from the code
>
>
> https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
>
> it looks like your user is null when trying to update the resources for
> the topology?  Did the supervisor and your topology recover when the
> supervisor was relaunched?  If not please file a bug JIRA and we can look
> into it.
>
>
> - Bobby
>
>
>
> On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya <
> sa...@haulmatic.com> wrote:
>
>
> Hi,
>
> We are using apache storm to analyze a GPS data stream by subscribing to
> rabbit mq message channel. We use apache storm 1.11.
>
> We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon
> EC2 instances. Also we have a local storm supervisor which is pointing to
> the remote nimbus in ec2 via the same zookeeper.
>
>
> When we run the topology as a local cluster or submit to the storm when
> only local supervisor is running ( we stop the remote supervisor instance),
> everything works as expected.
>
> Problem arise when we submit the topology to the remote supervisor
> (production like env). This keeps restarting the supervisor with below logs.
>
>
> *worker.log*
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e
> 19f051 6700
>
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating
> messaging context
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> executors
> 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex
> ecutor[2 2] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2
> 2]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-
> executor[3 3] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3
> 3]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor __acker:[1 1]
> 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1]
> [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1
> 1]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor __acker:[1 1]
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor rabbit-mq-gps-reader-spout:[6 6]
> 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6
> 6]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader
> -spout-executor[6 6] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor rabbit-mq-gps-reader-spout:[6 6]
> 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor __system:[-1 -1]
> 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1
> -1]-send-queue [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1]
> [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor __system:[-1 -1]
> 2017-08

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Bobby Evans
No you should not need to configure the users ahead of time.  In storm if 
security is turned off the "user" or "owner" of the topology is stored 
internally as the unix user running nimbus.  The fact that it is null indicates 
that there is some kind of a bug.  I am just trying to gauge how serious of a 
bug it is.  Does the supervisor recover after a little while or is it stuck in 
a bad state?


- Bobby


On Monday, August 7, 2017, 9:28:46 AM CDT, Martin Burian 
 wrote:

I am seeing exactly the same exception being thrown by supervisors in my 
topology after an update from storm 1.0.3 to 1.0.4.A quick search took me to 
http://storm.apache.org/releases/1.0.4/SECURITY.html where there are multiple 
references to some "users". Do I need to configure one of those? I Didn't have 
to do anything with users before.
Martin

po 7. 8. 2017 v 16:01 odesílatel Bobby Evans  napsal:

from the code
https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
it looks like your user is null when trying to update the resources for the 
topology?  Did the supervisor and your topology recover when the supervisor was 
relaunched?  If not please file a bug JIRA and we can look into it.


- Bobby


On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya 
 wrote:


Hi,

We are using apache storm to analyze a GPS data stream by subscribing to rabbit 
mq message channel. We use apache storm 1.11.

We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 
instances. Also we have a local storm supervisor which is pointing to the 
remote nimbus in ec2 via the same zookeeper. 





When we run the topology as a local cluster or submit to the storm when only 
local supervisor is running ( we stop the remote supervisor instance), 
everything works as expected.

Problem arise when we submit the topology to the remote supervisor (production 
like env). This keeps restarting the supervisor with below logs.




worker.log
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker 
GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 
6700
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating 
messaging context
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down 
executors
    2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 
2] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 
2]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- 
executor[3 3] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 
3]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __acker:[1 1]
    2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] 
Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 
1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor __acker:[1 1]
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 
6]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader 
-spout-executor[6 6] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __system:[-1 -1]
    2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 
-1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] 
[INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor __system:[-1 -1]
    2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor gps-data-logger:[5 5]
    2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 
5]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec utor[5 5] 
[INFO] Async loop interrupted!
    2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor gps-data-logger:[5

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Martin Burian
I am seeing exactly the same exception being thrown by supervisors in my
topology after an update from storm 1.0.3 to 1.0.4.
A quick search took me to
http://storm.apache.org/releases/1.0.4/SECURITY.html where there are
multiple references to some "users". Do I need to configure one of those? I
Didn't have to do anything with users before.

Martin

po 7. 8. 2017 v 16:01 odesílatel Bobby Evans  napsal:

> from the code
>
>
> https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
>
> it looks like your user is null when trying to update the resources for
> the topology?  Did the supervisor and your topology recover when the
> supervisor was relaunched?  If not please file a bug JIRA and we can look
> into it.
>
>
> - Bobby
>
>
>
> On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya <
> sa...@haulmatic.com> wrote:
>
>
> Hi,
>
> We are using apache storm to analyze a GPS data stream by subscribing to
> rabbit mq message channel. We use apache storm 1.11.
>
> We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon
> EC2 instances. Also we have a local storm supervisor which is pointing to
> the remote nimbus in ec2 via the same zookeeper.
>
>
> When we run the topology as a local cluster or submit to the storm when
> only local supervisor is running ( we stop the remote supervisor instance),
> everything works as expected.
>
> Problem arise when we submit the topology to the remote supervisor
> (production like env). This keeps restarting the supervisor with below logs.
>
>
> *worker.log*
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> worker GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e
> 19f051 6700
>
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating
> messaging context
> 2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
> executors
> 2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex
> ecutor[2 2] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2
> 2]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-analyzer:[2 2]
> 2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-
> executor[3 3] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3
> 3]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor fuel-data-save-to-db:[3 3]
> 2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor __acker:[1 1]
> 2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1]
> [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1
> 1]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor __acker:[1 1]
> 2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor rabbit-mq-gps-reader-spout:[6 6]
> 2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6
> 6]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader
> -spout-executor[6 6] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor rabbit-mq-gps-reader-spout:[6 6]
> 2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor __system:[-1 -1]
> 2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1
> -1]-send-queue [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1]
> [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor __system:[-1 -1]
> 2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor gps-data-logger:[5 5]
> 2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5
> 5]-send-queue [INFO] Async loop interrupted!
> 2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec
> utor[5 5] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down
> executor gps-data-logger:[5 5]
> 2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting
> down executor gps-data-devider:[4 4]
> 2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-exe
> cutor[4 4] [INFO] Async loop interrupted!
>
> 2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4
> 4]-send

Re: Storm workers get killed in the middle and supervisor restarts

2017-08-07 Thread Bobby Evans
from the code
https://github.com/apache/storm/blob/v1.1.1/storm-core/src/jvm/org/apache/storm/localizer/Localizer.java?utf8=%E2%9C%93#L332
it looks like your user is null when trying to update the resources for the 
topology?  Did the supervisor and your topology recover when the supervisor was 
relaunched?  If not please file a bug JIRA and we can look into it.


- Bobby


On Sunday, August 6, 2017, 11:34:19 PM CDT, Sahan Maldeniya 
 wrote:


Hi,

We are using apache storm to analyze a GPS data stream by subscribing to rabbit 
mq message channel. We use apache storm 1.11.

We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon EC2 
instances. Also we have a local storm supervisor which is pointing to the 
remote nimbus in ec2 via the same zookeeper. 





When we run the topology as a local cluster or submit to the storm when only 
local supervisor is running ( we stop the remote supervisor instance), 
everything works as expected.

Problem arise when we submit the topology to the remote supervisor (production 
like env). This keeps restarting the supervisor with below logs.




worker.log
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down worker 
GpsDataAnalyticsTopology-6-150 1922728 a383ada8-62a4-418c-9c7c-4e5a5e 19f051 
6700
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating 
messaging context
    2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down 
executors
    2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-ex ecutor[2 
2] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2 
2]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-analyzer:[2 2]
    2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db- 
executor[3 3] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3 
3]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor fuel-data-save-to-db:[3 3]
    2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __acker:[1 1]
    2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1] [INFO] 
Async loop interrupted!
    2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1 
1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor __acker:[1 1]
    2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6 
6]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.344 o.a.s.util Thread-10-rabbit-mq-gps-reader 
-spout-executor[6 6] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor rabbit-mq-gps-reader-spout:[6 6]
    2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor __system:[-1 -1]
    2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[- 1 
-1]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1] 
[INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor __system:[-1 -1]
    2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor gps-data-logger:[5 5]
    2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5 
5]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-exec utor[5 5] 
[INFO] Async loop interrupted!
    2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor gps-data-logger:[5 5]
    2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting down 
executor gps-data-devider:[4 4]
    2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-exe cutor[4 
4] [INFO] Async loop interrupted!
    2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4 
4]-send-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down 
executor gps-data-devider:[4 4]
    2017-08-05 08:50:27.351 o.a.s.d.worker Thread-19 [INFO] Shut down executors
    2017-08-05 08:50:27.353 o.a.s.d.worker Thread-19 [INFO] Shutting down 
transfer thread
    2017-08-05 08:50:27.353 o.a.s.util Thread-17-disruptor-worker-tra 
nsfer-queue [INFO] Async loop interrupted!
    2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down transfer 
thread
    2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut dow

Storm workers get killed in the middle and supervisor restarts

2017-08-06 Thread Sahan Maldeniya
Hi,

We are using apache storm to analyze a GPS data stream by subscribing to
rabbit mq message channel. We use apache storm 1.11.

We have deployed a zookeeper, 1 nimbus, 1 UI and a supervisor in 3 Amazon
EC2 instances. Also we have a local storm supervisor which is pointing to
the remote nimbus in ec2 via the same zookeeper.


When we run the topology as a local cluster or submit to the storm when
only local supervisor is running ( we stop the remote supervisor instance),
everything works as expected.

Problem arise when we submit the topology to the remote supervisor
(production like env). This keeps restarting the supervisor with below logs.


*worker.log*
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
worker GpsDataAnalyticsTopology-6-1501922728
a383ada8-62a4-418c-9c7c-4e5a5e19f051
6700
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Terminating
messaging context
2017-08-05 08:50:27.334 o.a.s.d.worker Thread-19 [INFO] Shutting down
executors
2017-08-05 08:50:27.334 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor fuel-data-analyzer:[2 2]
2017-08-05 08:50:27.335 o.a.s.util Thread-4-fuel-data-analyzer-executor[2
2] [INFO] Async loop interrupted!
2017-08-05 08:50:27.335 o.a.s.util Thread-3-disruptor-executor[2
2]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.341 o.a.s.d.executor Thread-19 [INFO] Shut down
executor fuel-data-analyzer:[2 2]
2017-08-05 08:50:27.342 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor fuel-data-save-to-db:[3 3]
2017-08-05 08:50:27.342 o.a.s.util Thread-6-fuel-data-save-to-db-executor[3
3] [INFO] Async loop interrupted!
2017-08-05 08:50:27.342 o.a.s.util Thread-5-disruptor-executor[3
3]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shut down
executor fuel-data-save-to-db:[3 3]
2017-08-05 08:50:27.343 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor __acker:[1 1]
2017-08-05 08:50:27.343 o.a.s.util Thread-8-__acker-executor[1 1]
[INFO] Async loop interrupted!
2017-08-05 08:50:27.343 o.a.s.util Thread-7-disruptor-executor[1
1]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shut down
executor __acker:[1 1]
2017-08-05 08:50:27.344 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor rabbit-mq-gps-reader-spout:[6 6]
2017-08-05 08:50:27.344 o.a.s.util Thread-9-disruptor-executor[6
6]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.344 o.a.s.util
Thread-10-rabbit-mq-gps-reader-spout-executor[6
6] [INFO] Async loop interrupted!
2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shut down
executor rabbit-mq-gps-reader-spout:[6 6]
2017-08-05 08:50:27.348 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor __system:[-1 -1]
2017-08-05 08:50:27.348 o.a.s.util Thread-11-disruptor-executor[-1
-1]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.349 o.a.s.util Thread-12-__system-executor[-1 -1]
[INFO] Async loop interrupted!
2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shut down
executor __system:[-1 -1]
2017-08-05 08:50:27.349 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor gps-data-logger:[5 5]
2017-08-05 08:50:27.349 o.a.s.util Thread-13-disruptor-executor[5
5]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.349 o.a.s.util Thread-14-gps-data-logger-executor[5
5] [INFO] Async loop interrupted!
2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down
executor gps-data-logger:[5 5]
2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shutting down
executor gps-data-devider:[4 4]
2017-08-05 08:50:27.350 o.a.s.util Thread-16-gps-data-devider-executor[4
4] [INFO] Async loop interrupted!
2017-08-05 08:50:27.350 o.a.s.util Thread-15-disruptor-executor[4
4]-send-queue [INFO] Async loop interrupted!
2017-08-05 08:50:27.350 o.a.s.d.executor Thread-19 [INFO] Shut down
executor gps-data-devider:[4 4]
2017-08-05 08:50:27.351 o.a.s.d.worker Thread-19 [INFO] Shut down
executors
2017-08-05 08:50:27.353 o.a.s.d.worker Thread-19 [INFO] Shutting down
transfer thread
2017-08-05 08:50:27.353 o.a.s.util
Thread-17-disruptor-worker-transfer-queue
[INFO] Async loop interrupted!
2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down
transfer thread
2017-08-05 08:50:27.354 o.a.s.d.worker Thread-19 [INFO] Shut down
backpressure thread
2017-08-05 08:50:27.355 o.a.s.d.worker Thread-19 [INFO] Shutting down
default resources
2017-08-05 08:50:27.356 o.a.s.d.worker Thread-19 [INFO] Shut down
default resources
2017-08-05 08:50:27.356 o.a.s.d.worker Thread-19 [INFO] Trigger any
worker shutdown hooks
2017-08-05 08:50:27.363 o.a.s.d.worker Thread-19 [INFO] Disconnecting
from storm cluster state context
2017-08-05 08:50:27.363 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl
Curator-Framework-0 [INFO] backg