[jira] [Comment Edited] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-20 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15593665#comment-15593665
 ] 

kasim edited comment on MESOS-6400 at 10/21/16 1:53 AM:


[~gilbert], there are two machines's logs, from 2016.9.26 to 2016.10.21

https://drive.google.com/open?id=0B1ULA2gXggVwOWQwLUxCYmdoWnc
https://drive.google.com/open?id=0B1ULA2gXggVwRmJCRmI1Qm1OOEU

1. Mesos and Marathon out of sync happen on someday in  2016.10.1 - 2016.10.7 
2. I restart Marathon in 2016.10.8



was (Author: mithril):
[~gilbert] There are two machines's logs, from 2016.9.26 to 2016.10.21

https://drive.google.com/open?id=0B1ULA2gXggVwOWQwLUxCYmdoWnc
https://drive.google.com/open?id=0B1ULA2gXggVwRmJCRmI1Qm1OOEU

1. Mesos and Marathon out of sync happen on someday in  2016.10.1 - 2016.10.7 
2. I restart Marathon in 2016.10.8


> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. Orphan Tasks still taked the resource, I have to delete them.
> 4. I find all Orphan Tasks are under framework 
> `ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-20 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15593665#comment-15593665
 ] 

kasim edited comment on MESOS-6400 at 10/21/16 1:53 AM:


[~gilbert] There are two machines's logs, from 2016.9.26 to 2016.10.21

https://drive.google.com/open?id=0B1ULA2gXggVwOWQwLUxCYmdoWnc
https://drive.google.com/open?id=0B1ULA2gXggVwRmJCRmI1Qm1OOEU

1. Mesos and Marathon out of sync happen on someday in  2016.10.1 - 2016.10.7 
2. I restart Marathon in 2016.10.8



was (Author: mithril):
There are two machines's logs, from 2016.9.26 to 2016.10.21

https://drive.google.com/open?id=0B1ULA2gXggVwOWQwLUxCYmdoWnc
https://drive.google.com/open?id=0B1ULA2gXggVwRmJCRmI1Qm1OOEU

1. Mesos and Marathon out of sync happen on someday in  2016.10.1 - 2016.10.7 
2. I restart Marathon in 2016.10.8


> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. Orphan Tasks still taked the resource, I have to delete them.
> 4. I find all Orphan Tasks are under framework 
> `ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-20 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15593665#comment-15593665
 ] 

kasim commented on MESOS-6400:
--

There are two machines's logs, from 2016.9.26 to 2016.10.21

https://drive.google.com/open?id=0B1ULA2gXggVwOWQwLUxCYmdoWnc
https://drive.google.com/open?id=0B1ULA2gXggVwRmJCRmI1Qm1OOEU

1. Mesos and Marathon out of sync happen on someday in  2016.10.1 - 2016.10.7 
2. I restart Marathon in 2016.10.8


> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. Orphan Tasks still taked the resource, I have to delete them.
> 4. I find all Orphan Tasks are under framework 
> `ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-18 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15587231#comment-15587231
 ] 

kasim commented on MESOS-6400:
--

I am using marathon 1.3.0-1.0.506.el7.

Yes, I can restarted marathon get a new framework id, and start some tasks(all 
duplicated of Orphan Tasks). And due to lack of resouce, it can not start all 
tasks. so I'd like to remove Orphan Tasks immediately, is there any way to do ?

> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. Orphan Tasks still taked the resource, I have to delete them.
> 4. I find all Orphan Tasks are under framework 
> `ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-17 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6400:
-
Description: 
The problem maybe cause by Mesos and Marathon out of sync
https://github.com/mesosphere/marathon/issues/616

When I found Orphan Tasks happen, I
1. restart marathon
2. marathon do not sync Orphan Tasks, but start new tasks.
3. Orphan Tasks still taked the resource, I have to delete them.
4. I find all Orphan Tasks are under framework 
`ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
`unregistered_frameworks`
{code}
{
"frameworks": [
.
],
"completed_frameworks": [ ],
"unregistered_frameworks": [
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-"
]
}
{code}

5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
, but get `No framework found with specified ID`


So I have no idea to delete Orphan Tasks


  was:
The problem maybe cause by Mesos and Marathon out of sync
https://github.com/mesosphere/marathon/issues/616

When I found Orphan Tasks happen, I
1. restart marathon
2. marathon do not sync Orphan Tasks, but start new tasks.
3. Orphan Tasks still taked the resource, I have to delete them.
4. all Orphan Tasks is under `ef169d8a-24fc-41d1-8b0d-c67718937a48-`
curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
`unregistered_frameworks`
{code}
{
"frameworks": [
.
],
"completed_frameworks": [ ],
"unregistered_frameworks": [
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-"
]
}
{code}

5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
, but get `No framework found with specified ID`


So I have no idea to delete Orphan Tasks



> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. Orphan Tasks still taked the resource, I have to delete them.
> 4. I find all Orphan Tasks are under framework 
> `ef169d8a-24fc-41d1-8b0d-c67718937a48-`,
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MESOS-6400) Not able to remove Orphan Tasks

2016-10-16 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6400:
-
Description: 
The problem maybe cause by Mesos and Marathon out of sync
https://github.com/mesosphere/marathon/issues/616

When I found Orphan Tasks happen, I
1. restart marathon
2. marathon do not sync Orphan Tasks, but start new tasks.
3. but Orphan Tasks still taked the resource, have to delete them.
4. all Orphan Tasks in under `ef169d8a-24fc-41d1-8b0d-c67718937a48-`
curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
`unregistered_frameworks`
{code}
{
"frameworks": [
.
],
"completed_frameworks": [ ],
"unregistered_frameworks": [
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-"
]
}
{code}

5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
, but get `No framework found with specified ID`


So I have no idea to delete Orphan Tasks


  was:
The problem maybe cause by Mesos and Marathon out of sync
https://github.com/mesosphere/marathon/issues/616

When I found Orphan Tasks happen, I
1. restart marathon
2. marathon do not sync Orphan Tasks, but start new tasks.
3. curl -XGET `http://c196:5050/master/frameworks` :
{code}
{
"frameworks": [
.
],
"completed_frameworks": [ ],
"unregistered_frameworks": [
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-",
"ef169d8a-24fc-41d1-8b0d-c67718937a48-"
]
}
{code}

all Orphan Tasks in under `ef169d8a-24fc-41d1-8b0d-c67718937a48-`
5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
, but get `No framework found with specified ID`


So I have no idea to delete Orphan Tasks



> Not able to remove Orphan Tasks
> ---
>
> Key: MESOS-6400
> URL: https://issues.apache.org/jira/browse/MESOS-6400
> Project: Mesos
>  Issue Type: Bug
>Affects Versions: 1.0.1
> Environment: centos 7 x64
>Reporter: kasim
>
> The problem maybe cause by Mesos and Marathon out of sync
> https://github.com/mesosphere/marathon/issues/616
> When I found Orphan Tasks happen, I
> 1. restart marathon
> 2. marathon do not sync Orphan Tasks, but start new tasks.
> 3. but Orphan Tasks still taked the resource, have to delete them.
> 4. all Orphan Tasks in under `ef169d8a-24fc-41d1-8b0d-c67718937a48-`
> curl -XGET `http://c196:5050/master/frameworks` shows that framework is 
> `unregistered_frameworks`
> {code}
> {
> "frameworks": [
> .
> ],
> "completed_frameworks": [ ],
> "unregistered_frameworks": [
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-",
> "ef169d8a-24fc-41d1-8b0d-c67718937a48-"
> ]
> }
> {code}
> 5.Try {code}curl -XPOST http://c196:5050/master/teardown -d 
> 'frameworkId=ef169d8a-24fc-41d1-8b0d-c67718937a48-' {code}
> , but get `No framework found with specified ID`
> So I have no idea to delete Orphan Tasks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15524701#comment-15524701
 ] 

kasim edited comment on MESOS-6248 at 9/27/16 1:21 AM:
---

Thanks, create soft link `ln -s /lib/systemd/systemd /sbin/init` work around 
this error. It is weird `/sbin/init` suddentlly disappeared.


was (Author: mithril):
Thanks, create soft link `ln -s /lib/systemd/systemd /sbin/init` work around 
this erorr. It is weird `/sbin/init` suddentlly disappeared.

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>Assignee: Jan Schlicht
> Fix For: 1.1.0
>
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597189882008970days
> Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
> logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
> Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
> InitGoogleLogging() is written to STDERR
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
> main.cpp:243] Build: 2016-08-26 23:06:27 by centos
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
> main.cpp:244] Version: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
> main.cpp:247] Git tag: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
> main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
> logging.cpp:194] INFO level logging started!
> Sep 26 

[jira] [Commented] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15524701#comment-15524701
 ] 

kasim commented on MESOS-6248:
--

Thanks, create soft link `ln -s /lib/systemd/systemd /sbin/init` work around 
this eror. It is weird `/sbin/init` suddentlly disappeared.

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>Assignee: Jan Schlicht
> Fix For: 1.1.0
>
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597189882008970days
> Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
> logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
> Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
> InitGoogleLogging() is written to STDERR
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
> main.cpp:243] Build: 2016-08-26 23:06:27 by centos
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
> main.cpp:244] Version: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
> main.cpp:247] Git tag: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
> main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
> logging.cpp:194] INFO level logging started!
> Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15524701#comment-15524701
 ] 

kasim edited comment on MESOS-6248 at 9/27/16 1:20 AM:
---

Thanks, create soft link `ln -s /lib/systemd/systemd /sbin/init` work around 
this erorr. It is weird `/sbin/init` suddentlly disappeared.


was (Author: mithril):
Thanks, create soft link `ln -s /lib/systemd/systemd /sbin/init` work around 
this eror. It is weird `/sbin/init` suddentlly disappeared.

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>Assignee: Jan Schlicht
> Fix For: 1.1.0
>
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597189882008970days
> Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
> logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
> Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
> InitGoogleLogging() is written to STDERR
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
> main.cpp:243] Build: 2016-08-26 23:06:27 by centos
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
> main.cpp:244] Version: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
> main.cpp:247] Git tag: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
> main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
> logging.cpp:194] INFO level logging started!
> Sep 26 

[jira] [Issue Comment Deleted] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6248:
-
Comment: was deleted

(was: {code}/sbin/init {code} not exists. What it should be ? Can I fix it 
manually?)

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>Assignee: Jan Schlicht
> Fix For: 1.1.0
>
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597189882008970days
> Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
> logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
> Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
> InitGoogleLogging() is written to STDERR
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
> main.cpp:243] Build: 2016-08-26 23:06:27 by centos
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
> main.cpp:244] Version: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
> main.cpp:247] Git tag: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
> main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
> logging.cpp:194] INFO level logging started!
> Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> Sep 26 15:51:42 zk2 mesos-slave[4475]: *** 

[jira] [Commented] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15524628#comment-15524628
 ] 

kasim commented on MESOS-6248:
--

{code}/sbin/init {code} not exists. What it should be ? Can I fix it manually?

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>Assignee: Jan Schlicht
> Fix For: 1.1.0
>
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597189882008970days
> Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
> logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
> Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
> InitGoogleLogging() is written to STDERR
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
> main.cpp:243] Build: 2016-08-26 23:06:27 by centos
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
> main.cpp:244] Version: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
> main.cpp:247] Git tag: 1.0.1
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
> main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
> Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
> logging.cpp:194] INFO level logging started!
> Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> Sep 26 15:51:42 zk2 mesos-slave[4475]: *** 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim edited comment on MESOS-6248 at 9/26/16 9:21 AM:
---

Same error, I found 

*service mesos-slave start* excute 
*logged /usr/sbin/mesos-master "${args[@]:-}"* 
in *function slave*  in */usr/bin/mesos-init-wrapper*.

I print args out and excute directly:
{code}
[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support
{code}

{code}
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
{code}


was (Author: mithril):
Same error, I found 

*service mesos-slave start* excute 
*logged /usr/sbin/mesos-master "${args[@]:-}"* in *function slave*  in 
*/usr/bin/mesos-init-wrapper*.

I print args out and excute directly:
{code}
[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support
{code}

{code}
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
{code}

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim edited comment on MESOS-6248 at 9/26/16 9:21 AM:
---

Same error, I found 

*service mesos-slave start* excute 
*logged /usr/sbin/mesos-master "${args[@]:-}"* in *function slave*  in 
*/usr/bin/mesos-init-wrapper*.

I print args out and excute directly:
{code}
[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support
{code}

{code}
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
{code}


was (Author: mithril):
Same error, I found 

{code}service mesos-slave start{code} excute 
{code}logged /usr/sbin/mesos-master "${args[@]:-}"{code} in {code}function 
slave{code}  in {code}/usr/bin/mesos-init-wrapper{code}.

I print args out and excute directly:
{code}
[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support
{code}

{code}
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
{code}

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim edited comment on MESOS-6248 at 9/26/16 9:19 AM:
---

Same error, I found 

{code}service mesos-slave start{code} excute 
{code}logged /usr/sbin/mesos-master "${args[@]:-}"{code} in {code}function 
slave{code}  in {code}/usr/bin/mesos-init-wrapper{code}.

I print args out and excute directly:
{code}
[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support
{code}

{code}
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
{code}


was (Author: mithril):
Same error, I found 

`service mesos-slave start` excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in `function slave`  in 
`/usr/bin/mesos-init-wrapper`.

I print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There occurs an error 
> {code}
> mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.
> {code}
> when I start `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> {code}
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 

[jira] [Updated] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6248:
-
Description: 
There occurs an error 
{code}
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
{code}
when I start `mesos-slave`.

What I do before this error happen:

- service docker stop and add a `registry-mirror`
- service docker start
- service mesos-slave start


Log as below:

{code}
Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594654510807141days
Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
slave.cpp:2376] Ignoring updating pid for framework 
ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597189882008970days
Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
InitGoogleLogging() is written to STDERR
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
main.cpp:243] Build: 2016-08-26 23:06:27 by centos
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
main.cpp:244] Version: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
main.cpp:247] Git tag: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
logging.cpp:194] INFO level logging started!
Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** Aborted at 1474876302 (unix time) 
try "date -d @1474876302" if you are using GNU date ***
Sep 26 15:51:42 zk2 mesos-slave[4475]: PC: @ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** SIGABRT (@0x117b) received by PID 
4475 (TID 0x7f79577b98c0) from PID 4475; stack trace: ***
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f79551d8100 (unknown)
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491fce8 __GI_abort
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f7954917566 
__assert_fail_base
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f7954917612 

[jira] [Updated] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6248:
-
Description: 
There occurs an error 
`mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.` 
when I start `mesos-slave`.

What I do before this error happen:

- service docker stop and add a `registry-mirror`
- service docker start
- service mesos-slave start


Log as below:

Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594654510807141days
Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
slave.cpp:2376] Ignoring updating pid for framework 
ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597189882008970days
Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
InitGoogleLogging() is written to STDERR
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
main.cpp:243] Build: 2016-08-26 23:06:27 by centos
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
main.cpp:244] Version: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
main.cpp:247] Git tag: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
logging.cpp:194] INFO level logging started!
Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** Aborted at 1474876302 (unix time) 
try "date -d @1474876302" if you are using GNU date ***
Sep 26 15:51:42 zk2 mesos-slave[4475]: PC: @ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** SIGABRT (@0x117b) received by PID 
4475 (TID 0x7f79577b98c0) from PID 4475; stack trace: ***
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f79551d8100 (unknown)
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491fce8 __GI_abort
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f7954917566 
__assert_fail_base
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f7954917612 
__GI___assert_fail
Sep 26 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim edited comment on MESOS-6248 at 9/26/16 9:14 AM:
---

Same error, I found 

`service mesos-slave start` excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in `function slave`  in 
`/usr/bin/mesos-init-wrapper`.

I copy the print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```


was (Author: mithril):
Same error, I found 

service mesos-slave start excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in function slave
in `/usr/bin/mesos-init-wrapper`.

I copy the print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There is an error `mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.` when I start 
> `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> 
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 

[jira] [Comment Edited] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim edited comment on MESOS-6248 at 9/26/16 9:14 AM:
---

Same error, I found 

`service mesos-slave start` excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in `function slave`  in 
`/usr/bin/mesos-init-wrapper`.

I print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```


was (Author: mithril):
Same error, I found 

`service mesos-slave start` excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in `function slave`  in 
`/usr/bin/mesos-init-wrapper`.

I copy the print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There is an error `mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.` when I start 
> `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> 
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 

[jira] [Commented] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15522536#comment-15522536
 ] 

kasim commented on MESOS-6248:
--

Same error, I found 

service mesos-slave start excute 
`logged /usr/sbin/mesos-master "${args[@]:-}"` in function slave
in `/usr/bin/mesos-init-wrapper`.

I copy the print args out and excute directly:

[root@zk2 mesos]# mesos-slave 
--master=zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos 
--log_dir=/var/log/mesos --containerizers=docker,mesos 
--executor_registration_timeout=10mins --hostname=10.142.55.196 
--ip=10.142.55.196 --work_dir=/var/lib/mesos --no-systemd_enable_support


```
WARNING: Logging before InitGoogleLogging() is written to STDERR
I0926 18:12:24.743870  3658 main.cpp:243] Build: 2016-08-26 23:06:27 by centos
I0926 18:12:24.744038  3658 main.cpp:244] Version: 1.0.1
I0926 18:12:24.744047  3658 main.cpp:247] Git tag: 1.0.1
I0926 18:12:24.744055  3658 main.cpp:251] Git SHA: 
3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
I0926 18:12:24.745198  3658 logging.cpp:194] INFO level logging started!
mesos-slave: ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
*** Aborted at 1474884744 (unix time) try "date -d @1474884744" if you are 
using GNU date ***
PC: @ 0x7f4cc2fc25f7 __GI_raise
*** SIGABRT (@0xe4a) received by PID 3658 (TID 0x7f4cc5e5d8c0) from PID 3658; 
stack trace: ***
@ 0x7f4cc387c100 (unknown)
@ 0x7f4cc2fc25f7 __GI_raise
@ 0x7f4cc2fc3ce8 __GI_abort
@ 0x7f4cc2fbb566 __assert_fail_base
@ 0x7f4cc2fbb612 __GI___assert_fail
@ 0x7f4cc45dfe4a Result<>::error()
@ 0x7f4cc4cde9a3 _ZZN7systemd6existsEvENKUlvE_clEv.isra.297
@ 0x7f4cc4cdf0bd systemd::exists()
@   0x41063c main
@ 0x7f4cc2faeb15 __libc_start_main
@   0x412099 (unknown)
Aborted
```

> mesos-slave cannot start , Assertion `isError()' failed
> ---
>
> Key: MESOS-6248
> URL: https://issues.apache.org/jira/browse/MESOS-6248
> Project: Mesos
>  Issue Type: Bug
> Environment: mesos 1.01 centos 7
>Reporter: kasim
>
> There is an error `mesos-slave: 
> ../../3rdparty/stout/include/stout/result.hpp:134: const string& 
> Result::error() const [with T = std::basic_string; std::string = 
> std::basic_string]: Assertion `isError()' failed.` when I start 
> `mesos-slave`.
> What I do before this error happen:
> - service docker stop and add a `registry-mirror`
> - service docker start
> - service mesos-slave start
> Log as below:
> 
> Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594670214119757days
> Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594654510807141days
> Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
> slave.cpp:2376] Ignoring updating pid for framework 
> ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
> Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
> slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
> 4.594667358972014days
> Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
> slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
> 4.597191309582847days
> Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
> slave.cpp:4591] Current 

[jira] [Created] (MESOS-6248) mesos-slave cannot start , Assertion `isError()' failed

2016-09-26 Thread kasim (JIRA)
kasim created MESOS-6248:


 Summary: mesos-slave cannot start , Assertion `isError()' failed
 Key: MESOS-6248
 URL: https://issues.apache.org/jira/browse/MESOS-6248
 Project: Mesos
  Issue Type: Bug
 Environment: mesos 1.01 centos 7
Reporter: kasim


There is an error `mesos-slave: 
../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.` when I start 
`mesos-slave`.

What I do before this error happen:

- service docker stop and add a `registry-mirror`
- service docker start
- service mesos-slave start


Log as below:

Sep 26 15:29:56 zk2 mesos-slave[5050]: I0926 15:29:56.421854  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:30:56 zk2 mesos-slave[5050]: I0926 15:30:56.422523  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594670214119757days
Sep 26 15:31:56 zk2 mesos-slave[5050]: I0926 15:31:56.423301  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594654510807141days
Sep 26 15:31:57 zk2 mesos-slave[5050]: W0926 15:31:57.067425  5070 
slave.cpp:2376] Ignoring updating pid for framework 
ef169d8a-24fc-41d1-8b0d-c67718937a48- because it does not exist
Sep 26 15:32:56 zk2 mesos-slave[5050]: I0926 15:32:56.424700  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:33:56 zk2 mesos-slave[5050]: I0926 15:33:56.425257  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:34:56 zk2 mesos-slave[5050]: I0926 15:34:56.425698  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:35:56 zk2 mesos-slave[5050]: I0926 15:35:56.426959  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:36:56 zk2 mesos-slave[5050]: I0926 15:36:56.428061  5070 
slave.cpp:4591] Current disk usage 24.36%. Max allowed age: 
4.594667358972014days
Sep 26 15:37:56 zk2 mesos-slave[5050]: I0926 15:37:56.429162  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:38:56 zk2 mesos-slave[5050]: I0926 15:38:56.429811  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:39:56 zk2 mesos-slave[5050]: I0926 15:39:56.430641  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:40:56 zk2 mesos-slave[5050]: I0926 15:40:56.431457  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:41:56 zk2 mesos-slave[5050]: I0926 15:41:56.432929  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:42:56 zk2 mesos-slave[5050]: I0926 15:42:56.433922  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:43:56 zk2 mesos-slave[5050]: I0926 15:43:56.434734  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597191309582847days
Sep 26 15:44:56 zk2 mesos-slave[5050]: I0926 15:44:56.435595  5070 
slave.cpp:4591] Current disk usage 24.33%. Max allowed age: 
4.597189882008970days
Sep 26 15:45:48 zk2 mesos-slave[5050]: W0926 15:44:56.435595  5050 
logging.cpp:91] RAW: Received signal SIGTERM from process 1 of user 0; exiting
Sep 26 15:51:42 zk2 mesos-slave[4475]: WARNING: Logging before 
InitGoogleLogging() is written to STDERR
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.072932  4475 
main.cpp:243] Build: 2016-08-26 23:06:27 by centos
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073128  4475 
main.cpp:244] Version: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073138  4475 
main.cpp:247] Git tag: 1.0.1
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.073144  4475 
main.cpp:251] Git SHA: 3611eb0b7eea8d144e9b2e840e0ba16f2f659ee3
Sep 26 15:51:42 zk2 mesos-slave[4475]: I0926 15:51:42.075127  4475 
logging.cpp:194] INFO level logging started!
Sep 26 15:51:42 zk2 mesos-slave[4475]: mesos-slave: 
../../3rdparty/stout/include/stout/result.hpp:134: const string& 
Result::error() const [with T = std::basic_string; std::string = 
std::basic_string]: Assertion `isError()' failed.
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** Aborted at 1474876302 (unix time) 
try "date -d @1474876302" if you are using GNU date ***
Sep 26 15:51:42 zk2 mesos-slave[4475]: PC: @ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]: *** SIGABRT (@0x117b) received by PID 
4475 (TID 0x7f79577b98c0) from PID 4475; stack trace: ***
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f79551d8100 (unknown)
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491e5f7 __GI_raise
Sep 26 15:51:42 zk2 mesos-slave[4475]:@ 0x7f795491fce8 __GI_abort
Sep 26 15:51:42 

[jira] [Comment Edited] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15505567#comment-15505567
 ] 

kasim edited comment on MESOS-6205 at 9/20/16 4:53 AM:
---

Thanks, empty work_dir works. But I don't understand how this situation happen. 

At first, I started only one master and zookeeper for test.

{code}
$ cat /etc/mesos/zk
zk://10.142.55.190:2181/mesos
{code}

The slave on same machine was able to connect master, but other couldn't.

So I tried to start three mesos-master and zookeepers to consist cluster, and 
change `/etc/mesos/zk` to

{code}
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
{code}, then got above error.


Is this mean I need clear wrok_dir everytime when adding a new mesos-master?


was (Author: mithril):
Thanks, empty work_dir works. But I don't understand how this situation happen. 

At first, I started only one master and zookeeper for test.

{code}
$ cat /etc/mesos/zk
zk://10.142.55.190:2181/mesos
{code}

The slave on same machine was able to connect master, but other couldn't.

So I tried to start three master to consist cluster, change `/etc/mesos/zk` to

{code}
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
{code}, then got above error.


Is this mean I need clear wrok_dir everytime when adding a new mesos-master?

> mesos-master can not found mesos-slave, and elect a new leader in a short 
> interval
> --
>
> Key: MESOS-6205
> URL: https://issues.apache.org/jira/browse/MESOS-6205
> Project: Mesos
>  Issue Type: Bug
>  Components: master
> Environment: ubuntu 12 x64, centos 6.5 x64, centos 7.2 x64
>Reporter: kasim
>
> I follow this 
> [doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
>  to setup mesos cluster.
> There are three vm(ubuntu 12, centos 6.5, centos 7.2).
> {code}
> $ cat /etc/hosts
> 10.142.55.190 zk1
> 10.142.55.196 zk2
> 10.142.55.202 zk3
> {code}
> config in each mathine:
> {code}
> $ cat /etc/mesos/zk
> zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
> {code}
> 
> After start zookeeper, mesos-master and mesos-slave in three vm, I can view 
> the mesos webui(10.142.55.190:5050), but agents count is 0.
> After a little time, mesos page get error:
> {code}
> Failed to connect to 10.142.55.190:5050!
> Retrying in 16 seconds... 
> {code}
> (I found that zookeeper would elect a new leader in a short interval)
> 
> mesos-master cmd:
> {code}
> mesos-master --agent_ping_timeout="15secs" 
> --agent_reregister_timeout="10mins" --allocation_interval="1secs" 
> --allocator="HierarchicalDRF" --authenticate_agents="false" 
> --authenticate_frameworks="false" --authenticate_http_frameworks="false" 
> --authenticate_http_readonly="false" --authenticate_http_readwrite="false" 
> --authenticators="crammd5" --authorizers="local" --framework_sorter="drf" 
> --help="false" --hostname_lookup="true" --http_authenticators="basic" 
> --initialize_driver_logging="true" --ip="10.142.55.190" 
> --log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
> --logging_level="INFO" --max_agent_ping_timeouts="5" 
> --max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
> --port="5050" --quiet="false" --quorum="2" 
> --recovery_agent_removal_limit="100%" --registry="replicated_log" 
> --registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
> --registry_strict="false" --root_submissions="true" --user_sorter="drf" 
> --version="false" --webui_dir="/usr/share/mesos/webui" 
> --work_dir="/var/lib/mesos" 
> --zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
> {code}
> mesos-slave cmd:
> {code}
> mesos-slave --appc_simple_discovery_uri_prefix="http://; 
> --appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
> --authenticate_http_readwrite="false" --authenticatee="crammd5" 
> --authentication_backoff_factor="1secs" --authorizer="local" 
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
> --cgroups_root="mesos" --container_disk_watch_interval="15secs" 
> --containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
> --docker="docker" --docker_kill_orphans="true" 
> --docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
> --docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
> --docker_store_dir="/tmp/mesos/store/docker" 
> --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
> --enforce_container_disk_quota="false" 
> --executor_registration_timeout="1mins" 
> 

[jira] [Comment Edited] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15505567#comment-15505567
 ] 

kasim edited comment on MESOS-6205 at 9/20/16 4:51 AM:
---

Thanks, empty work_dir works. But I don't understand how this situation happen. 

At first, I started only one master and zookeeper for test.

{code}
$ cat /etc/mesos/zk
zk://10.142.55.190:2181/mesos
{code}

The slave on same machine was able to connect master, but other couldn't.

So I tried to start three master to consist cluster, change `/etc/mesos/zk` to

{code}
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
{code}, then got above error.


Is this mean I need clear wrok_dir everytime when adding a new mesos-master?


was (Author: mithril):
Thanks, empty work_dir work. But I don't understand how this situation happen. 

At first, I started only one master and zookeeper for test.

{code}
$ cat /etc/mesos/zk
zk://10.142.55.190:2181/mesos
{code}

The slave on same machine was able to connect master, but other couldn't.

So I tried to start three master to consist cluster, change `/etc/mesos/zk` to

{code}
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
{code}, then got above error.


Is this mean I need clear wrok_dir everytime when adding a new mesos-master?

> mesos-master can not found mesos-slave, and elect a new leader in a short 
> interval
> --
>
> Key: MESOS-6205
> URL: https://issues.apache.org/jira/browse/MESOS-6205
> Project: Mesos
>  Issue Type: Bug
>  Components: master
> Environment: ubuntu 12 x64, centos 6.5 x64, centos 7.2 x64
>Reporter: kasim
>
> I follow this 
> [doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
>  to setup mesos cluster.
> There are three vm(ubuntu 12, centos 6.5, centos 7.2).
> {code}
> $ cat /etc/hosts
> 10.142.55.190 zk1
> 10.142.55.196 zk2
> 10.142.55.202 zk3
> {code}
> config in each mathine:
> {code}
> $ cat /etc/mesos/zk
> zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
> {code}
> 
> After start zookeeper, mesos-master and mesos-slave in three vm, I can view 
> the mesos webui(10.142.55.190:5050), but agents count is 0.
> After a little time, mesos page get error:
> {code}
> Failed to connect to 10.142.55.190:5050!
> Retrying in 16 seconds... 
> {code}
> (I found that zookeeper would elect a new leader in a short interval)
> 
> mesos-master cmd:
> {code}
> mesos-master --agent_ping_timeout="15secs" 
> --agent_reregister_timeout="10mins" --allocation_interval="1secs" 
> --allocator="HierarchicalDRF" --authenticate_agents="false" 
> --authenticate_frameworks="false" --authenticate_http_frameworks="false" 
> --authenticate_http_readonly="false" --authenticate_http_readwrite="false" 
> --authenticators="crammd5" --authorizers="local" --framework_sorter="drf" 
> --help="false" --hostname_lookup="true" --http_authenticators="basic" 
> --initialize_driver_logging="true" --ip="10.142.55.190" 
> --log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
> --logging_level="INFO" --max_agent_ping_timeouts="5" 
> --max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
> --port="5050" --quiet="false" --quorum="2" 
> --recovery_agent_removal_limit="100%" --registry="replicated_log" 
> --registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
> --registry_strict="false" --root_submissions="true" --user_sorter="drf" 
> --version="false" --webui_dir="/usr/share/mesos/webui" 
> --work_dir="/var/lib/mesos" 
> --zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
> {code}
> mesos-slave cmd:
> {code}
> mesos-slave --appc_simple_discovery_uri_prefix="http://; 
> --appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
> --authenticate_http_readwrite="false" --authenticatee="crammd5" 
> --authentication_backoff_factor="1secs" --authorizer="local" 
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
> --cgroups_root="mesos" --container_disk_watch_interval="15secs" 
> --containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
> --docker="docker" --docker_kill_orphans="true" 
> --docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
> --docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
> --docker_store_dir="/tmp/mesos/store/docker" 
> --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
> --enforce_container_disk_quota="false" 
> --executor_registration_timeout="1mins" 
> --executor_shutdown_grace_period="5secs" 
> 

[jira] [Commented] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

[ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15505567#comment-15505567
 ] 

kasim commented on MESOS-6205:
--

Thanks, empty work_dir work. But I don't understand how this situation happen. 

At first, I started only one master and zookeeper for test.

{code}
$ cat /etc/mesos/zk
zk://10.142.55.190:2181/mesos
{code}

The slave on same machine was able to connect master, but other couldn't.

So I tried to start three master to consist cluster, change `/etc/mesos/zk` to

{code}
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
{code}, then got above error.


Is this mean I need clear wrok_dir everytime when adding a new mesos-master?

> mesos-master can not found mesos-slave, and elect a new leader in a short 
> interval
> --
>
> Key: MESOS-6205
> URL: https://issues.apache.org/jira/browse/MESOS-6205
> Project: Mesos
>  Issue Type: Bug
>  Components: master
> Environment: ubuntu 12 x64, centos 6.5 x64, centos 7.2 x64
>Reporter: kasim
>
> I follow this 
> [doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
>  to setup mesos cluster.
> There are three vm(ubuntu 12, centos 6.5, centos 7.2).
> {code}
> $ cat /etc/hosts
> 10.142.55.190 zk1
> 10.142.55.196 zk2
> 10.142.55.202 zk3
> {code}
> config in each mathine:
> {code}
> $ cat /etc/mesos/zk
> zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos
> {code}
> 
> After start zookeeper, mesos-master and mesos-slave in three vm, I can view 
> the mesos webui(10.142.55.190:5050), but agents count is 0.
> After a little time, mesos page get error:
> {code}
> Failed to connect to 10.142.55.190:5050!
> Retrying in 16 seconds... 
> {code}
> (I found that zookeeper would elect a new leader in a short interval)
> 
> mesos-master cmd:
> {code}
> mesos-master --agent_ping_timeout="15secs" 
> --agent_reregister_timeout="10mins" --allocation_interval="1secs" 
> --allocator="HierarchicalDRF" --authenticate_agents="false" 
> --authenticate_frameworks="false" --authenticate_http_frameworks="false" 
> --authenticate_http_readonly="false" --authenticate_http_readwrite="false" 
> --authenticators="crammd5" --authorizers="local" --framework_sorter="drf" 
> --help="false" --hostname_lookup="true" --http_authenticators="basic" 
> --initialize_driver_logging="true" --ip="10.142.55.190" 
> --log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
> --logging_level="INFO" --max_agent_ping_timeouts="5" 
> --max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
> --port="5050" --quiet="false" --quorum="2" 
> --recovery_agent_removal_limit="100%" --registry="replicated_log" 
> --registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
> --registry_strict="false" --root_submissions="true" --user_sorter="drf" 
> --version="false" --webui_dir="/usr/share/mesos/webui" 
> --work_dir="/var/lib/mesos" 
> --zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
> {code}
> mesos-slave cmd:
> {code}
> mesos-slave --appc_simple_discovery_uri_prefix="http://; 
> --appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
> --authenticate_http_readwrite="false" --authenticatee="crammd5" 
> --authentication_backoff_factor="1secs" --authorizer="local" 
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
> --cgroups_root="mesos" --container_disk_watch_interval="15secs" 
> --containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
> --docker="docker" --docker_kill_orphans="true" 
> --docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
> --docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
> --docker_store_dir="/tmp/mesos/store/docker" 
> --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
> --enforce_container_disk_quota="false" 
> --executor_registration_timeout="1mins" 
> --executor_shutdown_grace_period="5secs" 
> --fetcher_cache_dir="/tmp/mesos/fetch" --fetcher_cache_size="2GB" 
> --frameworks_home="" --gc_delay="1weeks" --gc_disk_headroom="0.1" 
> --hadoop_home="" --help="false" --hostname="10.142.55.190" 
> --hostname_lookup="true" --http_authenticators="basic" 
> --http_command_executor="false" --image_provisioner_backend="copy" 
> --initialize_driver_logging="true" --ip="10.142.55.190" 
> --isolation="posix/cpu,posix/mem" --launcher="posix" 
> --launcher_dir="/usr/libexec/mesos" --log_dir="/var/log/mesos" 
> --logbufsecs="0" --logging_level="INFO" 
> 

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos


After start zookeeper, mesos-master and mesos-slave in three vm, I can view the 
mesos webui(10.142.55.190:5050), but agents count is 0.
After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


mesos-master cmd:
```
mesos-master --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate_agents="false" --authenticate_frameworks="false" 
--authenticate_http_frameworks="false" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticators="crammd5" 
--authorizers="local" --framework_sorter="drf" --help="false" 
--hostname_lookup="true" --http_authenticators="basic" 
--initialize_driver_logging="true" --ip="10.142.55.190" 
--log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" --max_agent_ping_timeouts="5" 
--max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
--port="5050" --quiet="false" --quorum="2" 
--recovery_agent_removal_limit="100%" --registry="replicated_log" 
--registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
--registry_strict="false" --root_submissions="true" --user_sorter="drf" 
--version="false" --webui_dir="/usr/share/mesos/webui" 
--work_dir="/var/lib/mesos" 
--zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
```


mesos-slave cmd:
```
mesos-slave --appc_simple_discovery_uri_prefix="http://; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticatee="crammd5" 
--authentication_backoff_factor="1secs" --authorizer="local" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
--docker="docker" --docker_kill_orphans="true" 
--docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
--docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
--docker_store_dir="/tmp/mesos/store/docker" 
--docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" --fetcher_cache_dir="/tmp/mesos/fetch" 
--fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname="10.142.55.190" --hostname_lookup="true" 
--http_authenticators="basic" --http_command_executor="false" 
--image_provisioner_backend="copy" --initialize_driver_logging="true" 
--ip="10.142.55.190" --isolation="posix/cpu,posix/mem" --launcher="posix" 
--launcher_dir="/usr/libexec/mesos" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" 
--master="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos" 
--oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
--perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" 
--quiet="false" --recover="reconnect" --recovery_timeout="15mins" 
--registration_backoff_factor="1secs" --revocable_cpu_low_priority="true" 
--sandbox_directory="/mnt/mesos/sandbox" --strict="true" --switch_user="true" 
--systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/var/lib/mesos"
```

When I run mesos-master from command-line, I got 

```
I0919 17:20:19.286264 17550 replica.cpp:673] Replica in VOTING status received 
a broadcasted recover request from (583)@10.142.55.202:5050
F0919 17:20:20.009371 17556 master.cpp:1536] Recovery failed: Failed to recover 
registrar: Failed to perform fetch within 1mins
*** Check failure stack trace: ***
@ 0x7f9db78458dd  google::LogMessage::Fail()
@ 0x7f9db784771d  google::LogMessage::SendToLog()
@ 0x7f9db78454cc  google::LogMessage::Flush()
@ 0x7f9db7848019  google::LogMessageFatal::~LogMessageFatal()
@ 0x7f9db6e2dbbc  mesos::internal::master::fail()
@ 0x7f9db6e75b20  

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos


After start zookeeper, mesos-master and mesos-slave in three vm, I can view the 
mesos webui(10.142.55.190:5050), but agents count is 0.
After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


mesos-master cmd:
```
mesos-master --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate_agents="false" --authenticate_frameworks="false" 
--authenticate_http_frameworks="false" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticators="crammd5" 
--authorizers="local" --framework_sorter="drf" --help="false" 
--hostname_lookup="true" --http_authenticators="basic" 
--initialize_driver_logging="true" --ip="10.142.55.190" 
--log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" --max_agent_ping_timeouts="5" 
--max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
--port="5050" --quiet="false" --quorum="2" 
--recovery_agent_removal_limit="100%" --registry="replicated_log" 
--registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
--registry_strict="false" --root_submissions="true" --user_sorter="drf" 
--version="false" --webui_dir="/usr/share/mesos/webui" 
--work_dir="/var/lib/mesos" 
--zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
```


mesos-slave cmd:
```
mesos-slave --appc_simple_discovery_uri_prefix="http://; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticatee="crammd5" 
--authentication_backoff_factor="1secs" --authorizer="local" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
--docker="docker" --docker_kill_orphans="true" 
--docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
--docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
--docker_store_dir="/tmp/mesos/store/docker" 
--docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" --fetcher_cache_dir="/tmp/mesos/fetch" 
--fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname="10.142.55.190" --hostname_lookup="true" 
--http_authenticators="basic" --http_command_executor="false" 
--image_provisioner_backend="copy" --initialize_driver_logging="true" 
--ip="10.142.55.190" --isolation="posix/cpu,posix/mem" --launcher="posix" 
--launcher_dir="/usr/libexec/mesos" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" 
--master="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos" 
--oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
--perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" 
--quiet="false" --recover="reconnect" --recovery_timeout="15mins" 
--registration_backoff_factor="1secs" --revocable_cpu_low_priority="true" 
--sandbox_directory="/mnt/mesos/sandbox" --strict="true" --switch_user="true" 
--systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/var/lib/mesos"
```

When I run mesos-master from command-line, I got 

```
I0919 17:20:19.286264 17550 replica.cpp:673] Replica in VOTING status received 
a broadcasted recover request from (583)@10.142.55.202:5050
F0919 17:20:20.009371 17556 master.cpp:1536] Recovery failed: Failed to recover 
registrar: Failed to perform fetch within 1mins
*** Check failure stack trace: ***
@ 0x7f9db78458dd  google::LogMessage::Fail()
@ 0x7f9db784771d  google::LogMessage::SendToLog()
@ 0x7f9db78454cc  google::LogMessage::Flush()
@ 0x7f9db7848019  google::LogMessageFatal::~LogMessageFatal()
@ 0x7f9db6e2dbbc  mesos::internal::master::fail()
@ 0x7f9db6e75b20  

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos


After start zookeeper, mesos-master and mesos-slave in three vm, I can view the 
mesos webui(10.142.55.190:5050), but agents count is 0.
After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


mesos-master cmd:
```
mesos-master --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate_agents="false" --authenticate_frameworks="false" 
--authenticate_http_frameworks="false" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticators="crammd5" 
--authorizers="local" --framework_sorter="drf" --help="false" 
--hostname_lookup="true" --http_authenticators="basic" 
--initialize_driver_logging="true" --ip="10.142.55.190" 
--log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" --max_agent_ping_timeouts="5" 
--max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
--port="5050" --quiet="false" --quorum="2" 
--recovery_agent_removal_limit="100%" --registry="replicated_log" 
--registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
--registry_strict="false" --root_submissions="true" --user_sorter="drf" 
--version="false" --webui_dir="/usr/share/mesos/webui" 
--work_dir="/var/lib/mesos" 
--zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
```


mesos-slave cmd:
```
mesos-slave --appc_simple_discovery_uri_prefix="http://; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticatee="crammd5" 
--authentication_backoff_factor="1secs" --authorizer="local" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
--docker="docker" --docker_kill_orphans="true" 
--docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
--docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
--docker_store_dir="/tmp/mesos/store/docker" 
--docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" --fetcher_cache_dir="/tmp/mesos/fetch" 
--fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname="10.142.55.190" --hostname_lookup="true" 
--http_authenticators="basic" --http_command_executor="false" 
--image_provisioner_backend="copy" --initialize_driver_logging="true" 
--ip="10.142.55.190" --isolation="posix/cpu,posix/mem" --launcher="posix" 
--launcher_dir="/usr/libexec/mesos" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" 
--master="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos" 
--oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
--perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" 
--quiet="false" --recover="reconnect" --recovery_timeout="15mins" 
--registration_backoff_factor="1secs" --revocable_cpu_low_priority="true" 
--sandbox_directory="/mnt/mesos/sandbox" --strict="true" --switch_user="true" 
--systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/var/lib/mesos"
```

When I run mesos-master from command-line, I got 

```
I0919 17:20:19.286264 17550 replica.cpp:673] Replica in VOTING status received 
a broadcasted recover request from (583)@10.142.55.202:5050
F0919 17:20:20.009371 17556 master.cpp:1536] Recovery failed: Failed to recover 
registrar: Failed to perform fetch within 1mins
*** Check failure stack trace: ***
@ 0x7f9db78458dd  google::LogMessage::Fail()
@ 0x7f9db784771d  google::LogMessage::SendToLog()
@ 0x7f9db78454cc  google::LogMessage::Flush()
@ 0x7f9db7848019  google::LogMessageFatal::~LogMessageFatal()
@ 0x7f9db6e2dbbc  mesos::internal::master::fail()
@ 0x7f9db6e75b20  

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos

After start zookeeper, mesos-master and mesos-slave in three vm, I can view the 
mesos webui(10.142.55.190:5050), but agents count is 0.

After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)

mesos-master cmd:
```
mesos-master --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate_agents="false" --authenticate_frameworks="false" 
--authenticate_http_frameworks="false" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticators="crammd5" 
--authorizers="local" --framework_sorter="drf" --help="false" 
--hostname_lookup="true" --http_authenticators="basic" 
--initialize_driver_logging="true" --ip="10.142.55.190" 
--log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" --max_agent_ping_timeouts="5" 
--max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
--port="5050" --quiet="false" --quorum="2" 
--recovery_agent_removal_limit="100%" --registry="replicated_log" 
--registry_fetch_timeout="1mins" --registry_store_timeout="20secs" 
--registry_strict="false" --root_submissions="true" --user_sorter="drf" 
--version="false" --webui_dir="/usr/share/mesos/webui" 
--work_dir="/var/lib/mesos" 
--zk="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos"
```


mesos-slave cmd:
```
mesos-slave --appc_simple_discovery_uri_prefix="http://; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="false" 
--authenticate_http_readwrite="false" --authenticatee="crammd5" 
--authentication_backoff_factor="1secs" --authorizer="local" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" --default_role="*" --disk_watch_interval="1mins" 
--docker="docker" --docker_kill_orphans="true" 
--docker_registry="https://registry-1.docker.io; --docker_remove_delay="6hrs" 
--docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" 
--docker_store_dir="/tmp/mesos/store/docker" 
--docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" --fetcher_cache_dir="/tmp/mesos/fetch" 
--fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname="10.142.55.190" --hostname_lookup="true" 
--http_authenticators="basic" --http_command_executor="false" 
--image_provisioner_backend="copy" --initialize_driver_logging="true" 
--ip="10.142.55.190" --isolation="posix/cpu,posix/mem" --launcher="posix" 
--launcher_dir="/usr/libexec/mesos" --log_dir="/var/log/mesos" --logbufsecs="0" 
--logging_level="INFO" 
--master="zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos" 
--oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
--perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" 
--quiet="false" --recover="reconnect" --recovery_timeout="15mins" 
--registration_backoff_factor="1secs" --revocable_cpu_low_priority="true" 
--sandbox_directory="/mnt/mesos/sandbox" --strict="true" --switch_user="true" 
--systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/var/lib/mesos"
```

When I run mesos-master from command-line, I got 

```
I0919 17:20:19.286264 17550 replica.cpp:673] Replica in VOTING status received 
a broadcasted recover request from (583)@10.142.55.202:5050
F0919 17:20:20.009371 17556 master.cpp:1536] Recovery failed: Failed to recover 
registrar: Failed to perform fetch within 1mins
*** Check failure stack trace: ***
@ 0x7f9db78458dd  google::LogMessage::Fail()
@ 0x7f9db784771d  google::LogMessage::SendToLog()
@ 0x7f9db78454cc  google::LogMessage::Flush()
@ 0x7f9db7848019  google::LogMessageFatal::~LogMessageFatal()
@ 0x7f9db6e2dbbc  mesos::internal::master::fail()
@ 0x7f9db6e75b20  

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos

After start zookeeper, mesos-master and mesos-slave in three vm, I can view the 
mesos webui(10.142.55.190:5050), but agents count is 0.

After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


master info log:

I0919 15:54:59.677438 13281 http.cpp:2022] Redirecting request for 
/master/state?jsonp=angular.callbacks._1x to the leading master zk3
I0919 15:55:00.098667 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (768)@10.142.55.202:5050
I0919 15:55:00.385279 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (185)@10.142.55.196:5050
I0919 15:55:00.79 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (771)@10.142.55.202:5050
I0919 15:55:01.347291 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (188)@10.142.55.196:5050
I0919 15:55:01.597682 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (774)@10.142.55.202:5050
I0919 15:55:02.257159 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (191)@10.142.55.196:5050
I0919 15:55:02.370692 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (777)@10.142.55.202:5050
I0919 15:55:03.205920 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (780)@10.142.55.202:5050
I0919 15:55:03.260007 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (194)@10.142.55.196:5050
I0919 15:55:03.929611 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (783)@10.142.55.202:5050
I0919 15:55:04.033308 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (197)@10.142.55.196:5050
I0919 15:55:04.591275 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (200)@10.142.55.196:5050
I0919 15:55:04.608211 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (786)@10.142.55.202:5050
I0919 15:55:05.184682 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (789)@10.142.55.202:5050
I0919 15:55:05.268277 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (203)@10.142.55.196:5050
I0919 15:55:05.775377 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (206)@10.142.55.196:5050
I0919 15:55:05.916445 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (792)@10.142.55.202:5050
I0919 15:55:06.744927 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (209)@10.142.55.196:5050
I0919 15:55:07.378521 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (5)@10.142.55.202:5050
I0919 15:55:07.393311 13285 network.hpp:430] ZooKeeper group memberships 
changed
I0919 15:55:07.393427 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000709' in ZooKeeper
I0919 15:55:07.393985 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000711' in ZooKeeper
I0919 15:55:07.394394 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000714' in ZooKeeper
I0919 15:55:07.394843 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000715' in ZooKeeper
I0919 15:55:07.395418 13285 network.hpp:478] ZooKeeper group PIDs: { 
log-replica(1)@10.142.55.190:5050, log-replica(1)@10.142.55.196:5050, 
log-replica(1)@10.142.55.202:5050 }
I0919 15:55:08.178272 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (14)@10.142.55.202:5050
I0919 15:55:09.059562 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (21)@10.142.55.202:5050
I0919 15:55:09.700711 13286 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (24)@10.142.55.202:5050
I0919 15:55:09.742185 13287 http.cpp:381] HTTP GET for 

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos

After I start zookeeper, mesos-master and mesos-slave in three vm, I can view 
the mesos webui(10.142.55.190:5050). I found agents count is 0.

After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


master info log:

I0919 15:54:59.677438 13281 http.cpp:2022] Redirecting request for 
/master/state?jsonp=angular.callbacks._1x to the leading master zk3
I0919 15:55:00.098667 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (768)@10.142.55.202:5050
I0919 15:55:00.385279 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (185)@10.142.55.196:5050
I0919 15:55:00.79 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (771)@10.142.55.202:5050
I0919 15:55:01.347291 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (188)@10.142.55.196:5050
I0919 15:55:01.597682 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (774)@10.142.55.202:5050
I0919 15:55:02.257159 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (191)@10.142.55.196:5050
I0919 15:55:02.370692 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (777)@10.142.55.202:5050
I0919 15:55:03.205920 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (780)@10.142.55.202:5050
I0919 15:55:03.260007 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (194)@10.142.55.196:5050
I0919 15:55:03.929611 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (783)@10.142.55.202:5050
I0919 15:55:04.033308 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (197)@10.142.55.196:5050
I0919 15:55:04.591275 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (200)@10.142.55.196:5050
I0919 15:55:04.608211 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (786)@10.142.55.202:5050
I0919 15:55:05.184682 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (789)@10.142.55.202:5050
I0919 15:55:05.268277 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (203)@10.142.55.196:5050
I0919 15:55:05.775377 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (206)@10.142.55.196:5050
I0919 15:55:05.916445 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (792)@10.142.55.202:5050
I0919 15:55:06.744927 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (209)@10.142.55.196:5050
I0919 15:55:07.378521 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (5)@10.142.55.202:5050
I0919 15:55:07.393311 13285 network.hpp:430] ZooKeeper group memberships 
changed
I0919 15:55:07.393427 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000709' in ZooKeeper
I0919 15:55:07.393985 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000711' in ZooKeeper
I0919 15:55:07.394394 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000714' in ZooKeeper
I0919 15:55:07.394843 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000715' in ZooKeeper
I0919 15:55:07.395418 13285 network.hpp:478] ZooKeeper group PIDs: { 
log-replica(1)@10.142.55.190:5050, log-replica(1)@10.142.55.196:5050, 
log-replica(1)@10.142.55.202:5050 }
I0919 15:55:08.178272 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (14)@10.142.55.202:5050
I0919 15:55:09.059562 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (21)@10.142.55.202:5050
I0919 15:55:09.700711 13286 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (24)@10.142.55.202:5050
I0919 15:55:09.742185 13287 http.cpp:381] HTTP GET 

[jira] [Updated] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)

 [ 
https://issues.apache.org/jira/browse/MESOS-6205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kasim updated MESOS-6205:
-
Description: 
I follow this 
[doc][https://open.mesosphere.com/getting-started/install/#verifying-installation]
 to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos

After I start zookeeper, mesos-master and mesos-slave in three vm, I can view 
the mesos webui(10.142.55.190:5050). I found agents count is 0.

After a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


master info log:

I0919 15:54:59.677438 13281 http.cpp:2022] Redirecting request for 
/master/state?jsonp=angular.callbacks._1x to the leading master zk3
I0919 15:55:00.098667 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (768)@10.142.55.202:5050
I0919 15:55:00.385279 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (185)@10.142.55.196:5050
I0919 15:55:00.79 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (771)@10.142.55.202:5050
I0919 15:55:01.347291 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (188)@10.142.55.196:5050
I0919 15:55:01.597682 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (774)@10.142.55.202:5050
I0919 15:55:02.257159 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (191)@10.142.55.196:5050
I0919 15:55:02.370692 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (777)@10.142.55.202:5050
I0919 15:55:03.205920 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (780)@10.142.55.202:5050
I0919 15:55:03.260007 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (194)@10.142.55.196:5050
I0919 15:55:03.929611 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (783)@10.142.55.202:5050
I0919 15:55:04.033308 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (197)@10.142.55.196:5050
I0919 15:55:04.591275 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (200)@10.142.55.196:5050
I0919 15:55:04.608211 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (786)@10.142.55.202:5050
I0919 15:55:05.184682 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (789)@10.142.55.202:5050
I0919 15:55:05.268277 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (203)@10.142.55.196:5050
I0919 15:55:05.775377 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (206)@10.142.55.196:5050
I0919 15:55:05.916445 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (792)@10.142.55.202:5050
I0919 15:55:06.744927 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (209)@10.142.55.196:5050
I0919 15:55:07.378521 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (5)@10.142.55.202:5050
I0919 15:55:07.393311 13285 network.hpp:430] ZooKeeper group memberships 
changed
I0919 15:55:07.393427 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000709' in ZooKeeper
I0919 15:55:07.393985 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000711' in ZooKeeper
I0919 15:55:07.394394 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000714' in ZooKeeper
I0919 15:55:07.394843 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000715' in ZooKeeper
I0919 15:55:07.395418 13285 network.hpp:478] ZooKeeper group PIDs: { 
log-replica(1)@10.142.55.190:5050, log-replica(1)@10.142.55.196:5050, 
log-replica(1)@10.142.55.202:5050 }
I0919 15:55:08.178272 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (14)@10.142.55.202:5050
I0919 15:55:09.059562 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (21)@10.142.55.202:5050
I0919 15:55:09.700711 13286 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (24)@10.142.55.202:5050
I0919 15:55:09.742185 13287 http.cpp:381] HTTP GET 

[jira] [Created] (MESOS-6205) mesos-master can not found mesos-slave, and elect a new leader in a short interval

2016-09-19 Thread kasim (JIRA)
kasim created MESOS-6205:


 Summary: mesos-master can not found mesos-slave, and elect a new 
leader in a short interval
 Key: MESOS-6205
 URL: https://issues.apache.org/jira/browse/MESOS-6205
 Project: Mesos
  Issue Type: Bug
  Components: master
 Environment: ubuntu 12 x64, centos 6.5 x64, centos 7.2 x64
Reporter: kasim


I follow this [doc][1] to setup mesos cluster.

There are three vm(ubuntu 12, centos 6.5, centos 7.2).

$ cat /etc/hosts
10.142.55.190 zk1
10.142.55.196 zk2
10.142.55.202 zk3

config in each mathine:

$ cat /etc/mesos/zk
zk://10.142.55.190:2181,10.142.55.196:2181,10.142.55.202:2181/mesos

After I start mesos-master in three vm, I can view the mesos 
webui(10.142.55.190:5050),  but after a little time, mesos page get error:

Failed to connect to 10.142.55.190:5050!
Retrying in 16 seconds... 
(I found that zookeeper would elect a new leader in a short interval)


master info log:

I0919 15:54:59.677438 13281 http.cpp:2022] Redirecting request for 
/master/state?jsonp=angular.callbacks._1x to the leading master zk3
I0919 15:55:00.098667 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (768)@10.142.55.202:5050
I0919 15:55:00.385279 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (185)@10.142.55.196:5050
I0919 15:55:00.79 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (771)@10.142.55.202:5050
I0919 15:55:01.347291 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (188)@10.142.55.196:5050
I0919 15:55:01.597682 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (774)@10.142.55.202:5050
I0919 15:55:02.257159 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (191)@10.142.55.196:5050
I0919 15:55:02.370692 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (777)@10.142.55.202:5050
I0919 15:55:03.205920 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (780)@10.142.55.202:5050
I0919 15:55:03.260007 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (194)@10.142.55.196:5050
I0919 15:55:03.929611 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (783)@10.142.55.202:5050
I0919 15:55:04.033308 13287 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (197)@10.142.55.196:5050
I0919 15:55:04.591275 13284 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (200)@10.142.55.196:5050
I0919 15:55:04.608211 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (786)@10.142.55.202:5050
I0919 15:55:05.184682 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (789)@10.142.55.202:5050
I0919 15:55:05.268277 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (203)@10.142.55.196:5050
I0919 15:55:05.775377 13281 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (206)@10.142.55.196:5050
I0919 15:55:05.916445 13285 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (792)@10.142.55.202:5050
I0919 15:55:06.744927 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (209)@10.142.55.196:5050
I0919 15:55:07.378521 13283 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (5)@10.142.55.202:5050
I0919 15:55:07.393311 13285 network.hpp:430] ZooKeeper group memberships 
changed
I0919 15:55:07.393427 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000709' in ZooKeeper
I0919 15:55:07.393985 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000711' in ZooKeeper
I0919 15:55:07.394394 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000714' in ZooKeeper
I0919 15:55:07.394843 13285 group.cpp:706] Trying to get 
'/mesos/log_replicas/000715' in ZooKeeper
I0919 15:55:07.395418 13285 network.hpp:478] ZooKeeper group PIDs: { 
log-replica(1)@10.142.55.190:5050, log-replica(1)@10.142.55.196:5050, 
log-replica(1)@10.142.55.202:5050 }
I0919 15:55:08.178272 13280 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (14)@10.142.55.202:5050
I0919 15:55:09.059562 13282 replica.cpp:673] Replica in VOTING status 
received a broadcasted recover request from (21)@10.142.55.202:5050
I0919 15:55:09.700711 13286 replica.cpp:673] Replica in VOTING status