Thanks for the info. I did find the "Removed agent" line as you suspected,
but not much else in logging looked promising. I opened a JIRA to track
from here on out https://issues.apache.org/jira/browse/MESOS-9555.

On Tue, Feb 5, 2019 at 2:03 PM Joseph Wu <jos...@mesosphere.io> wrote:

> From the stack, it looks like the master is attempting to remove an agent
> from the master's in-memory state.  In the master's logs you should find a
> line shortly before the exit, like:
>
> <timestamp> master.cpp:nnnn] Removed agent <ID of agent>: <reason>
>
> The agent's ID should at least give you some pointer to which agent is
> causing the problem.  Feel free to create a JIRA (
> https://issues.apache.org/jira/) with any information you can glean.
> This particular type of failure, a CHECK-failure, means some invariant has
> been violated and usually means we missed a corner case.
>
> On Tue, Feb 5, 2019 at 12:04 PM Jeff Pollard <jeff.poll...@gmail.com>
> wrote:
>
>> We recently upgraded our Mesos  cluster from version 1.3 to 1.5, and
>> since then have been getting periodic master crashes due to this error:
>>
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]: F0205 15:53:57.385118
>> 8434 hierarchical.cpp:2630] Check failed:
>> reservationScalarQuantities.contains(role)
>>
>> Full stack trace is at the end of this email. When the master fails, we
>> automatically restart it and it rejoins the cluster just fine. I did some
>> initial searching and was unable to find any existing bug reports or other
>> people experiencing this issue. We run a cluster of 3 masters, and see
>> crashes on all 3 instances.
>>
>> Hope to get some guidance on what is going on and/or where to start
>> looking for more information.
>>
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e9170a7d  google::LogMessage::Fail()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e9172830  google::LogMessage::SendToLog()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e9170663  google::LogMessage::Flush()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e9173259  google::LogMessageFatal::~LogMessageFatal()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e8443cbd
>> mesos::internal::master::allocator::internal::HierarchicalAllocatorProcess::untrackReservations()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e8448fcd
>> mesos::internal::master::allocator::internal::HierarchicalAllocatorProcess::removeSlave()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e90c4f11  process::ProcessBase::consume()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e90dea4a  process::ProcessManager::resume()
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e90e25d6
>> _ZNSt6thread5_ImplISt12_Bind_simpleIFZN7process14ProcessManager12init_threadsEvEUlvE_vEEE6_M_runEv
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e6700c80  (unknown)
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e5f136ba  start_thread
>> Feb  5 15:53:57 ip-10-0-16-140 mesos-master[8414]:     @
>>  0x7f87e5c4941d  (unknown)
>>
>

Reply via email to