[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Attachment: RoleTest.ImplicitRoleRegister.txt > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann >Assignee: haosdent > Labels: mesosphere, tests > Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, > CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log, > RoleTest.ImplicitRoleRegister.txt, > flaky-containerizer-pid-namespace-backward.txt, > flaky-containerizer-pid-namespace-forward.txt > > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) > * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Attachment: flaky-containerizer-pid-namespace-forward.txt flaky-containerizer-pid-namespace-backward.txt > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann >Assignee: haosdent > Labels: mesosphere, tests > Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, > CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log, > flaky-containerizer-pid-namespace-backward.txt, > flaky-containerizer-pid-namespace-forward.txt > > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) > * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Joseph Wu updated MESOS-6180: - Attachment: FetchAndStoreAndStoreAndFetch.log CGROUPS_ROOT_PidNamespaceForward.log CGROUPS_ROOT_PidNamespaceBackward.log I've attached logs for the three at the top of the list. > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann >Assignee: haosdent > Labels: mesosphere, tests > Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, > CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log > > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) > * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Description: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166) See the linked JIRAs noted above for individual tickets addressing a couple of these. was: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) See the linked JIRAs noted above for individual tickets addressing a couple of these. > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann > Labels: mesosphere, tests > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) > * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Description: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) See the linked JIRAs noted above for individual tickets addressing a couple of these. was: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) See the linked JIRAs noted above for individual tickets addressing a couple of these. > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann > Labels: mesosphere, tests > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Description: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) See the linked JIRAs noted above for individual tickets addressing a couple of these. was: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) See the linked JIRAs for individual tickets addressing a couple of these. > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann > Labels: mesosphere, tests > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > See the linked JIRAs noted above for individual tickets addressing a couple > of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early
[ https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Mann updated MESOS-6180: - Description: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch * RoleTest.ImplicitRoleRegister * SlaveRecoveryTest/0.MultipleFrameworks * SlaveRecoveryTest/0.ReconcileShutdownFramework * SlaveTest.ContainerizerUsageFailure * MesosSchedulerDriverTest.ExplicitAcknowledgements * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) See the linked JIRAs for individual tickets addressing a couple of these. was: Following the merging of a large patch chain, it was noticed on our internal CI that several tests had become flaky, with a similar pattern in the failures: the tests fail early when a future times out. Often, this occurs when a test cluster is being spun up and one of the offer futures times out. This has been observed in the following tests: * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward *MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward *ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch *RoleTest.ImplicitRoleRegister *SlaveRecoveryTest/0.MultipleFrameworks *SlaveRecoveryTest/0.ReconcileShutdownFramework *SlaveTest.ContainerizerUsageFailure *MesosSchedulerDriverTest.ExplicitAcknowledgements *SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) See the linked JIRAs for individual tickets addressing a couple of these. > Several tests are flaky, with futures timing out early > -- > > Key: MESOS-6180 > URL: https://issues.apache.org/jira/browse/MESOS-6180 > Project: Mesos > Issue Type: Bug > Components: tests >Reporter: Greg Mann > Labels: mesosphere, tests > > Following the merging of a large patch chain, it was noticed on our internal > CI that several tests had become flaky, with a similar pattern in the > failures: the tests fail early when a future times out. Often, this occurs > when a test cluster is being spun up and one of the offer futures times out. > This has been observed in the following tests: > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward > * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward > * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch > * RoleTest.ImplicitRoleRegister > * SlaveRecoveryTest/0.MultipleFrameworks > * SlaveRecoveryTest/0.ReconcileShutdownFramework > * SlaveTest.ContainerizerUsageFailure > * MesosSchedulerDriverTest.ExplicitAcknowledgements > * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164) > See the linked JIRAs for individual tickets addressing a couple of these. -- This message was sent by Atlassian JIRA (v6.3.4#6332)