[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-20 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Attachment: RoleTest.ImplicitRoleRegister.txt

> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>Assignee: haosdent
>  Labels: mesosphere, tests
> Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, 
> CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log, 
> RoleTest.ImplicitRoleRegister.txt, 
> flaky-containerizer-pid-namespace-backward.txt, 
> flaky-containerizer-pid-namespace-forward.txt
>
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
> * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-16 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Attachment: flaky-containerizer-pid-namespace-forward.txt
flaky-containerizer-pid-namespace-backward.txt

> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>Assignee: haosdent
>  Labels: mesosphere, tests
> Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, 
> CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log, 
> flaky-containerizer-pid-namespace-backward.txt, 
> flaky-containerizer-pid-namespace-forward.txt
>
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
> * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-15 Thread Joseph Wu (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joseph Wu updated MESOS-6180:
-
Attachment: FetchAndStoreAndStoreAndFetch.log
CGROUPS_ROOT_PidNamespaceForward.log
CGROUPS_ROOT_PidNamespaceBackward.log

I've attached logs for the three at the top of the list.

> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>Assignee: haosdent
>  Labels: mesosphere, tests
> Attachments: CGROUPS_ROOT_PidNamespaceBackward.log, 
> CGROUPS_ROOT_PidNamespaceForward.log, FetchAndStoreAndStoreAndFetch.log
>
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
> * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-15 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Description: 
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
* ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
* SlaveTest.CommandTaskWithKillPolicy (MESOS-6166)

See the linked JIRAs noted above for individual tickets addressing a couple of 
these.

  was:
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
* ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)

See the linked JIRAs noted above for individual tickets addressing a couple of 
these.


> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>  Labels: mesosphere, tests
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
> * SlaveTest.CommandTaskWithKillPolicy (MESOS-6166)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-15 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Description: 
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
* ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)

See the linked JIRAs noted above for individual tickets addressing a couple of 
these.

  was:
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)

See the linked JIRAs noted above for individual tickets addressing a couple of 
these.


> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>  Labels: mesosphere, tests
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> * ResourceOffersTest.ResourcesGetReofferedAfterTaskInfoError (MESOS-6165)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-15 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Description: 
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)

See the linked JIRAs noted above for individual tickets addressing a couple of 
these.

  was:
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)

See the linked JIRAs for individual tickets addressing a couple of these.


> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>  Labels: mesosphere, tests
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> See the linked JIRAs noted above for individual tickets addressing a couple 
> of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6180) Several tests are flaky, with futures timing out early

2016-09-15 Thread Greg Mann (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Mann updated MESOS-6180:
-
Description: 
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
* ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
* RoleTest.ImplicitRoleRegister
* SlaveRecoveryTest/0.MultipleFrameworks
* SlaveRecoveryTest/0.ReconcileShutdownFramework
* SlaveTest.ContainerizerUsageFailure
* MesosSchedulerDriverTest.ExplicitAcknowledgements
* SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)

See the linked JIRAs for individual tickets addressing a couple of these.

  was:
Following the merging of a large patch chain, it was noticed on our internal CI 
that several tests had become flaky, with a similar pattern in the failures: 
the tests fail early when a future times out. Often, this occurs when a test 
cluster is being spun up and one of the offer futures times out. This has been 
observed in the following tests:

* MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
*MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
*ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
*RoleTest.ImplicitRoleRegister
*SlaveRecoveryTest/0.MultipleFrameworks
*SlaveRecoveryTest/0.ReconcileShutdownFramework
*SlaveTest.ContainerizerUsageFailure
*MesosSchedulerDriverTest.ExplicitAcknowledgements
*SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)

See the linked JIRAs for individual tickets addressing a couple of these.


> Several tests are flaky, with futures timing out early
> --
>
> Key: MESOS-6180
> URL: https://issues.apache.org/jira/browse/MESOS-6180
> Project: Mesos
>  Issue Type: Bug
>  Components: tests
>Reporter: Greg Mann
>  Labels: mesosphere, tests
>
> Following the merging of a large patch chain, it was noticed on our internal 
> CI that several tests had become flaky, with a similar pattern in the 
> failures: the tests fail early when a future times out. Often, this occurs 
> when a test cluster is being spun up and one of the offer futures times out. 
> This has been observed in the following tests:
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceForward
> * MesosContainerizerSlaveRecoveryTest.CGROUPS_ROOT_PidNamespaceBackward
> * ZooKeeperStateTest.FetchAndStoreAndStoreAndFetch
> * RoleTest.ImplicitRoleRegister
> * SlaveRecoveryTest/0.MultipleFrameworks
> * SlaveRecoveryTest/0.ReconcileShutdownFramework
> * SlaveTest.ContainerizerUsageFailure
> * MesosSchedulerDriverTest.ExplicitAcknowledgements
> * SlaveRecoveryTest/0.ReconnectHTTPExecutor (MESOS-6164)
> See the linked JIRAs for individual tickets addressing a couple of these.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)