John,
I was referring to the following log message...isn't that the right one?
Sep 26 18:11:58 machine62 aurora-scheduler[24743]: I0926 18:11:58.795
[Thread-814,
MesosCallbackHandler$MesosCallbackHandlerImpl:180] Registered with ID
value: "4ca9aa06-3214-4d2c-a678-0832e2f84d17-0000"

On Tue, Sep 26, 2017 at 3:30 PM, John Sirois <[email protected]> wrote:

> > ... but it succeeded(see logs at the end) ...
>
> The underlying c++ code in the scheduler driver successfully connected to
> the leading master socket:
> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926 18:11:37.231549 24868
> scheduler.cpp:361] Connected with the master at
> http://10.163.25.45:5050/master/api/v1/scheduler
>
> <http://10.163.25.45:5050/master/api/v1/scheduler>This is not the same as
> a framework registration call being successfully executed against the newly
> connected master.
> You need to be careful about what you derive from the logs just based on a
> reading of the words. Generally you'll need to look carefully / grep
> sourcecode to be sure you are mentally modelling the code flows correctly.
> It certainly gets tricky.
>
> On Tue, Sep 26, 2017 at 4:14 PM, Mohit Jaggi <[email protected]> wrote:
>
>> Hmm....so it is indeed the mesos registration, but it succeeded(see logs
>> at the end). Looking at the code, I see that  com.google.common.eventbus
>> is used as pub-sub mechanism to link the registration (first log message)
>>  to the  registrationAcked flag...and this flag is not being set for
>> 10mins (?) otherwise the registration timeout handler will not print the
>> second log message.
>>
>> delayedActions.onRegistrationTimeout(
>>     () -> {
>>       if (!registrationAcked.get()) {
>>         LOG.error(
>>             "Framework has not been registered within the tolerated delay.");
>>         stateMachine.transition(State.DEAD);
>>       }
>>     });
>>
>> Sep 26 18:11:58 machine62 aurora-scheduler[24743]: I0926 18:11:58.795
>> [Thread-814, MesosCallbackHandler$MesosCallbackHandlerImpl:180]
>> Registered with ID value: "4ca9aa06-3214-4d2c-a678-0832e2f84d17-0000"
>>
>> ...
>>
>> ...
>>
>> Sep 26 18:21:37 machine62 aurora-scheduler[24743]: E0926 18:21:37.205
>> [Lifecycle-0, SchedulerLifecycle$4:235] Framework has not been registered
>> within the tolerated delay.
>>
>> On Tue, Sep 26, 2017 at 2:34 PM, John Sirois <[email protected]>
>> wrote:
>>
>>>
>>>
>>> On Tue, Sep 26, 2017 at 3:31 PM, Mohit Jaggi <[email protected]>
>>> wrote:
>>>
>>>> Actually I wonder if the "register" here is for Aurora to register with
>>>> mesos as a framework...I was assuming it refered to registering with ZK as
>>>> a leader.
>>>>
>>>
>>> Bingo.
>>>
>>>
>>>>
>>>> On Tue, Sep 26, 2017 at 2:27 PM, Renan DelValle Rueda <
>>>> [email protected]> wrote:
>>>>
>>>>> I've had this, or something very similar, happen before. It's an issue
>>>>> with Aurora and ZK. Election is based upon ZK, so if writing down who the
>>>>> leader is to the ZK server path fails, or if ZK is unable to reach quorum
>>>>> on the write, the election will fail. Sometimes this might manifest itself
>>>>> in weird ways, such as two aurora schedulers believing they are leaders. 
>>>>> If
>>>>> you could tell us a little bit about your ZK set up we might be able to
>>>>> narrow down the issue. Also, Aurora version and whether you are using
>>>>> Curator or the commons library will help as well.
>>>>>
>>>>> On Tue, Sep 26, 2017 at 2:02 PM, Mohit Jaggi <[email protected]>
>>>>> wrote:
>>>>>
>>>>>> Hmm..it seems machine62 became a leader but could not "register" as
>>>>>> leader. Not sure what that means. My naive assumption is that "becoming
>>>>>> leader" and "registering as leader" are "atomic".
>>>>>>
>>>>>> ------- grep on SchedulerLifecycle -----
>>>>>> aurora-scheduler.log:Sep 26 18:11:33 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:11:33.158 [LeaderSelector-0,
>>>>>> StateMachine$Builder:389] SchedulerLifecycle state machine transition
>>>>>> STORAGE_PREPARED -> LEADER_AWAITING_REGISTRATION
>>>>>> aurora-scheduler.log:Sep 26 18:11:33 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:11:33.159 [LeaderSelector-0,
>>>>>> SchedulerLifecycle$4:224] Elected as leading scheduler!
>>>>>> aurora-scheduler.log:Sep 26 18:11:37 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:11:37.204 [LeaderSelector-0,
>>>>>> SchedulerLifecycle$DefaultDelayedActions:163] Giving up on
>>>>>> registration in (10, mins)
>>>>>> aurora-scheduler.log:Sep 26 18:21:37 machine62
>>>>>> aurora-scheduler[24743]: E0926 18:21:37.205 [Lifecycle-0,
>>>>>> SchedulerLifecycle$4:235] Framework has not been registered within the
>>>>>> tolerated delay.
>>>>>> aurora-scheduler.log:Sep 26 18:21:37 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:21:37.205 [Lifecycle-0,
>>>>>> StateMachine$Builder:389] SchedulerLifecycle state machine transition
>>>>>> LEADER_AWAITING_REGISTRATION -> DEAD
>>>>>> aurora-scheduler.log:Sep 26 18:21:37 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:21:37.215 [Lifecycle-0,
>>>>>> StateMachine$Builder:389] SchedulerLifecycle state machine transition 
>>>>>> DEAD
>>>>>> -> DEAD
>>>>>> aurora-scheduler.log:Sep 26 18:21:37 machine62
>>>>>> aurora-scheduler[24743]: I0926 18:21:37.215 [Lifecycle-0,
>>>>>> SchedulerLifecycle$6:275] Shutdown already invoked, ignoring extra call.
>>>>>> aurora-scheduler.log:Sep 26 18:22:05 machine62
>>>>>> aurora-scheduler[54502]: I0926 18:22:05.681 [main,
>>>>>> StateMachine$Builder:389] SchedulerLifecycle state machine transition 
>>>>>> IDLE
>>>>>> -> PREPARING_STORAGE
>>>>>> aurora-scheduler.log:Sep 26 18:22:06 machine62
>>>>>> aurora-scheduler[54502]: I0926 18:22:06.396 [main,
>>>>>> StateMachine$Builder:389] SchedulerLifecycle state machine transition
>>>>>> PREPARING_STORAGE -> STORAGE_PREPARED
>>>>>>
>>>>>>
>>>>>> ------ connecting to mesos -----
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.211750 24871 group.cpp:757] Found non-sequence node 
>>>>>> 'log_replicas'
>>>>>> at '/mesos' in ZooKeeper
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.211817 24871 detector.cpp:152] Detected a new leader: 
>>>>>> (id='1506')
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.211917 24871 group.cpp:699] Trying to get
>>>>>> '/mesos/json.info_0000001506' in ZooKeeper
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.216063 24871 zookeeper.cpp:262] A new leading master (UPID=
>>>>>> [email protected]:5050) is detected
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.216162 24871 scheduler.cpp:470] New master detected at
>>>>>> [email protected]:5050
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.217772 24871 scheduler.cpp:479] Waiting for 12.81503ms before
>>>>>> initiating a re-(connection) attempt with the master
>>>>>> Sep 26 18:11:37 machine62 aurora-scheduler[24743]: I0926
>>>>>> 18:11:37.231549 24868 scheduler.cpp:361] Connected with the master at
>>>>>> http://10.163.25.45:5050/master/api/v1/scheduler
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Sep 26, 2017 at 1:24 PM, Bill Farner <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> Is there a reason a non-leading scheduler will talk to Mesos
>>>>>>>
>>>>>>>
>>>>>>> No, there is not a legitimate reason.  Did this occur for an
>>>>>>> extended period of time?  Do you have logs from the scheduler indicating
>>>>>>> that it lost ZK leadership and subsequently interacted with mesos?
>>>>>>>
>>>>>>> On Tue, Sep 26, 2017 at 1:02 PM, Mohit Jaggi <[email protected]>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Fellows,
>>>>>>>> While examining Aurora log files, I noticed a condition where a
>>>>>>>> scheduler was talking to Mesos but it was not showing up as a leader in
>>>>>>>> Zookeeper. It ultimately restarted itself and another scheduler became 
>>>>>>>> the
>>>>>>>> leader.
>>>>>>>> Is there a reason a non-leading scheduler will talk to Mesos?
>>>>>>>>
>>>>>>>> Mohit.
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to