[ 
https://issues.apache.org/jira/browse/MESOS-7466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joseph Wu updated MESOS-7466:
-----------------------------
      Docs Text:   (was: I submitted an application group in marathon, and then 
I deleted the application group because there was some problem with the 
definition of the application group, and then I submitted the application group 
again via the marathon api. After multiple deletions and multiple submissions, 
I found some problems (this is the third time I encountered the problem), I 
deleted the application group in the marathon UI, found in the `deployments` 
there are some content, and can not be deleted, suggesting that `error 
destroying null: app '/ Null' does not exist `; in the mesos ui task status is 
`running`; docker container status is `up`. Then, I restart mesos-master, 
mesos-slave, marathon and zookeeper, the result is the same.

I suggested that I clear the marathon in the zookeeper and then I did; marathon 
was back to normal, just like reinstalling it. But the tasks I want to delete 
are still in mesos ui and the state is `running`.

After that, I cleared the meso in zookeeper, and the result is the same. The 
task I want to delete is still in mesos ui. But I have to delete the task in 
the docker does not exist, I do not know why.

The follow is my config info:

```
$ curl -sSL http://172.30.30.4:5050/version | python -m json.tool
{
    "build_date": "2017-04-12 16:39:09",
    "build_time": 1492015149.0,
    "build_user": "centos",
    "git_sha": "de306b5786de3c221bae1457c6f2ccaeb38eef9f",
    "git_tag": "1.2.0",
    "version": "1.2.0"
}

$ curl -sSL http://172.30.30.4:8080/v2/info | python -m json.tool
{
    ......
    "name": "marathon",
    "version": "1.4.2",
    ......
}

$ docker version
Client:
 Version:      1.12.5
 API version:  1.24
 Go version:   go1.6.4
 Git commit:   7392c3b
 Built:        Fri Dec 16 02:23:59 2016
 OS/Arch:      linux/amd64

Server:
 Version:      1.12.5
 API version:  1.24
 Go version:   go1.6.4
 Git commit:   7392c3b
 Built:        Fri Dec 16 02:23:59 2016
 OS/Arch:      linux/amd64

$ docker info
Containers: 2
 Running: 1
 Paused: 0
 Stopped: 1
Images: 16
Server Version: 1.12.5
Storage Driver: devicemapper
 Pool Name: docker-253:0-403269431-pool
 Pool Blocksize: 65.54 kB
 Base Device Size: 10.74 GB
 Backing Filesystem: xfs
 Data file: /dev/loop0
 Metadata file: /dev/loop1
 Data Space Used: 4.915 GB
 Data Space Total: 107.4 GB
 Data Space Available: 94.26 GB
 Metadata Space Used: 7.115 MB
 Metadata Space Total: 2.147 GB
 Metadata Space Available: 2.14 GB
 Thin Pool Minimum Free Space: 10.74 GB
 Udev Sync Supported: true
 Deferred Removal Enabled: false
 Deferred Deletion Enabled: false
 Deferred Deleted Device Count: 0
 Data loop file: /var/lib/docker/devicemapper/devicemapper/data
 WARNING: Usage of loopback devices is strongly discouraged for production use. 
Use `--storage-opt dm.thinpooldev` to specify a custom block storage device.
 Metadata loop file: /var/lib/docker/devicemapper/devicemapper/metadata
 Library Version: 1.02.135-RHEL7 (2016-09-28)
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
 Volume: local
 Network: calico null bridge host overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options: seccomp
Kernel Version: 3.10.0-514.2.2.el7.x86_64
Operating System: CentOS Linux 7 (Core)
OSType: linux
Architecture: x86_64
CPUs: 8
Total Memory: 15.51 GiB
Name: slave1
ID: 4VRS:RPSC:ASAX:MTRJ:TOGC:6RFL:6RJS:J4MK:3VSN:JZO2:Q2FB:LJEW
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Cluster Store: etcd://172.30.30.5:2379
Cluster Advertise: 172.30.30.5:2375
Insecure Registries:
 172.30.30.8:80
 127.0.0.0/8

$ /usr/sbin/mesos-master \
    --zk=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
    --port=5050 \
    --log_dir=/var/log/mesos \
    --cluster=cp_cluster \
    --hostname=172.30.30.4 \
    --quorum=1 \
    --work_dir=/var/lib/mesos

$ /usr/sbin/mesos-slave \
    --master=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
    --log_dir=/var/log/mesos \
    --containerizers=docker,mesos \
    --executor_registration_timeout=5mins \
    --hostname=172.30.30.5 \
    --ip=172.30.30.5 \
    --isolation=filesystem/linux,docker/runtime,network/cni \
    --network_cni_config_dir=/var/lib/mesos/cni/config \
    --network_cni_plugins_dir=/var/lib/mesos/cni/plugins \
    --work_dir=/var/lib/mesos

$ curl -sSL http://172.30.30.12:5050/master/tasks | python -m json.tool | grep 
"\"id\""
            "id": "monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-tools_prometheus.c6bdfccc-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-tools_alertmanager.c3bfff0b-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_node-exporter.0e44c1c5-3232-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_cadvisor.0e455e07-3232-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_mesos-exporter.0e43fe73-3232-11e7-bd32-024222b481ef",
            "id": "syslog.e0645b43-3238-11e7-90c2-02423620ccdc",
            "id": "syslog.e72bbd15-3238-11e7-90c2-02423620ccdc",
            "id": "syslog.a0b162ff-316f-11e7-bd32-024222b481ef",
            "id": "mesos-dns.1286663d-30ae-11e7-bd32-024222b481ef",
            "id": "mesos-dns.12c20fae-30ae-11e7-bd32-024222b481ef",
            "id": "mesos-dns.1286181c-30ae-11e7-bd32-024222b481ef",

# slave1
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
e05468eaedc2        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node

# slave2
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
7fcbb3f2fa61        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node

# slave3
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
c69eba5bc322        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node
```

`monitor-***` tasks are I want to delete, but these tasks do not exist in the 
docker (`docker ps -a` is not the task).

The follow is `monitor-***`tasks log info(`stdout`):

```
......
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Re-registered docker executor on 172.30.30.6
Re-registered docker executor on 172.30.30.6
```

Now it's `stderr`:

```
I0506 23:26:45.211597 17482 exec.cpp:162] Version: 1.2.0
I0506 23:26:45.224807 17489 exec.cpp:237] Executor registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 23:26:45.310137 17486 docker.cpp:850] Running docker -H 
unix:///var/run/docker.sock run --cpu-shares 512 --memory 1073741824 --env-file 
/tmp/qT2brz -v /etc/localtime:/etc/localtime:ro -v 
/var/lib/mesos/slaves/e79deb05-5f20-48ca-9de9-a9610504e040-S1/frameworks/1e68ea0f-0f0b-4f14-8e2e-ab10169ee5f3-0000/executors/monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef/runs/7674214b-a609-47e6-a51a-7129616e8494:/mnt/mesos/sandbox
 --net calico1 --log-driver=syslog 
--log-opt=syslog-address=tcp://172.30.30.11:514 --log-opt=tag=grafana 
--label=MESOS_TASK_ID=monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
 --name 
mesos-e79deb05-5f20-48ca-9de9-a9610504e040-S1.7674214b-a609-47e6-a51a-7129616e8494
 grafana/grafana
I0506 16:57:13.687083 17489 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 16:57:14.359922 17486 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 16:57:14.366843 17488 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:40.419311 17485 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:41.629977 17485 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:41.658910 17485 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:02:20.707939 17485 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:03:43.669370 17485 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:03:43.698617 17485 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
```

Now i want to know how to remove these tasks, and where are these tasks 
running? (I do not think run in the mesos container because I specified 
`"type": "DOCKER"`)

There is a slave also thrown the following exception:

```
Message from syslogd@slave2 at May  6 08:27:35 ...
 kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1

Message from syslogd@slave2 at May  6 08:27:45 ...
 kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1
```)
         Labels:   (was: github-import)
    Description: 
I submitted an application group in marathon, and then I deleted the 
application group because there was some problem with the definition of the 
application group, and then I submitted the application group again via the 
marathon api. After multiple deletions and multiple submissions, I found some 
problems (this is the third time I encountered the problem), I deleted the 
application group in the marathon UI, found in the `deployments` there are some 
content, and can not be deleted, suggesting that `error destroying null: app '/ 
Null' does not exist `; in the mesos ui task status is `running`; docker 
container status is `up`. Then, I restart mesos-master, mesos-slave, marathon 
and zookeeper, the result is the same.

I suggested that I clear the marathon in the zookeeper and then I did; marathon 
was back to normal, just like reinstalling it. But the tasks I want to delete 
are still in mesos ui and the state is `running`.

After that, I cleared the meso in zookeeper, and the result is the same. The 
task I want to delete is still in mesos ui. But I have to delete the task in 
the docker does not exist, I do not know why.

The follow is my config info:

{noformat}
$ curl -sSL http://172.30.30.4:5050/version | python -m json.tool
{
    "build_date": "2017-04-12 16:39:09",
    "build_time": 1492015149.0,
    "build_user": "centos",
    "git_sha": "de306b5786de3c221bae1457c6f2ccaeb38eef9f",
    "git_tag": "1.2.0",
    "version": "1.2.0"
}

$ curl -sSL http://172.30.30.4:8080/v2/info | python -m json.tool
{
    ......
    "name": "marathon",
    "version": "1.4.2",
    ......
}

$ docker version
Client:
 Version:      1.12.5
 API version:  1.24
 Go version:   go1.6.4
 Git commit:   7392c3b
 Built:        Fri Dec 16 02:23:59 2016
 OS/Arch:      linux/amd64

Server:
 Version:      1.12.5
 API version:  1.24
 Go version:   go1.6.4
 Git commit:   7392c3b
 Built:        Fri Dec 16 02:23:59 2016
 OS/Arch:      linux/amd64

$ docker info
Containers: 2
 Running: 1
 Paused: 0
 Stopped: 1
Images: 16
Server Version: 1.12.5
Storage Driver: devicemapper
 Pool Name: docker-253:0-403269431-pool
 Pool Blocksize: 65.54 kB
 Base Device Size: 10.74 GB
 Backing Filesystem: xfs
 Data file: /dev/loop0
 Metadata file: /dev/loop1
 Data Space Used: 4.915 GB
 Data Space Total: 107.4 GB
 Data Space Available: 94.26 GB
 Metadata Space Used: 7.115 MB
 Metadata Space Total: 2.147 GB
 Metadata Space Available: 2.14 GB
 Thin Pool Minimum Free Space: 10.74 GB
 Udev Sync Supported: true
 Deferred Removal Enabled: false
 Deferred Deletion Enabled: false
 Deferred Deleted Device Count: 0
 Data loop file: /var/lib/docker/devicemapper/devicemapper/data
 WARNING: Usage of loopback devices is strongly discouraged for production use. 
Use `--storage-opt dm.thinpooldev` to specify a custom block storage device.
 Metadata loop file: /var/lib/docker/devicemapper/devicemapper/metadata
 Library Version: 1.02.135-RHEL7 (2016-09-28)
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
 Volume: local
 Network: calico null bridge host overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options: seccomp
Kernel Version: 3.10.0-514.2.2.el7.x86_64
Operating System: CentOS Linux 7 (Core)
OSType: linux
Architecture: x86_64
CPUs: 8
Total Memory: 15.51 GiB
Name: slave1
ID: 4VRS:RPSC:ASAX:MTRJ:TOGC:6RFL:6RJS:J4MK:3VSN:JZO2:Q2FB:LJEW
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Cluster Store: etcd://172.30.30.5:2379
Cluster Advertise: 172.30.30.5:2375
Insecure Registries:
 172.30.30.8:80
 127.0.0.0/8

$ /usr/sbin/mesos-master \
    --zk=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
    --port=5050 \
    --log_dir=/var/log/mesos \
    --cluster=cp_cluster \
    --hostname=172.30.30.4 \
    --quorum=1 \
    --work_dir=/var/lib/mesos

$ /usr/sbin/mesos-slave \
    --master=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
    --log_dir=/var/log/mesos \
    --containerizers=docker,mesos \
    --executor_registration_timeout=5mins \
    --hostname=172.30.30.5 \
    --ip=172.30.30.5 \
    --isolation=filesystem/linux,docker/runtime,network/cni \
    --network_cni_config_dir=/var/lib/mesos/cni/config \
    --network_cni_plugins_dir=/var/lib/mesos/cni/plugins \
    --work_dir=/var/lib/mesos

$ curl -sSL http://172.30.30.12:5050/master/tasks | python -m json.tool | grep 
"\"id\""
            "id": "monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-tools_prometheus.c6bdfccc-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-tools_alertmanager.c3bfff0b-3233-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_node-exporter.0e44c1c5-3232-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_cadvisor.0e455e07-3232-11e7-bd32-024222b481ef",
            "id": 
"monitor-exporter_mesos-exporter.0e43fe73-3232-11e7-bd32-024222b481ef",
            "id": "syslog.e0645b43-3238-11e7-90c2-02423620ccdc",
            "id": "syslog.e72bbd15-3238-11e7-90c2-02423620ccdc",
            "id": "syslog.a0b162ff-316f-11e7-bd32-024222b481ef",
            "id": "mesos-dns.1286663d-30ae-11e7-bd32-024222b481ef",
            "id": "mesos-dns.12c20fae-30ae-11e7-bd32-024222b481ef",
            "id": "mesos-dns.1286181c-30ae-11e7-bd32-024222b481ef",

# slave1
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
e05468eaedc2        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node

# slave2
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
7fcbb3f2fa61        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node

# slave3
$ docker ps -a
CONTAINER ID        IMAGE                        COMMAND             CREATED    
         STATUS              PORTS               NAMES
c69eba5bc322        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
ago         Up 4 hours                              calico-node
{noformat}

`monitor-***` tasks are I want to delete, but these tasks do not exist in the 
docker (`docker ps -a` is not the task).

The follow is `monitor-***`tasks log info(`stdout`):

{noformat}
......
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Received killTask for task 
monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
Re-registered docker executor on 172.30.30.6
Re-registered docker executor on 172.30.30.6
{noformat}

Now it's `stderr`:

{noformat}
I0506 23:26:45.211597 17482 exec.cpp:162] Version: 1.2.0
I0506 23:26:45.224807 17489 exec.cpp:237] Executor registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 23:26:45.310137 17486 docker.cpp:850] Running docker -H 
unix:///var/run/docker.sock run --cpu-shares 512 --memory 1073741824 --env-file 
/tmp/qT2brz -v /etc/localtime:/etc/localtime:ro -v 
/var/lib/mesos/slaves/e79deb05-5f20-48ca-9de9-a9610504e040-S1/frameworks/1e68ea0f-0f0b-4f14-8e2e-ab10169ee5f3-0000/executors/monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef/runs/7674214b-a609-47e6-a51a-7129616e8494:/mnt/mesos/sandbox
 --net calico1 --log-driver=syslog 
--log-opt=syslog-address=tcp://172.30.30.11:514 --log-opt=tag=grafana 
--label=MESOS_TASK_ID=monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
 --name 
mesos-e79deb05-5f20-48ca-9de9-a9610504e040-S1.7674214b-a609-47e6-a51a-7129616e8494
 grafana/grafana
I0506 16:57:13.687083 17489 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 16:57:14.359922 17486 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 16:57:14.366843 17488 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:40.419311 17485 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:41.629977 17485 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 17:45:41.658910 17485 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:02:20.707939 17485 exec.cpp:488] Agent exited, but framework has 
checkpointing enabled. Waiting 15mins to reconnect with agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:03:43.669370 17485 exec.cpp:283] Received reconnect request from agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
I0506 18:03:43.698617 17485 exec.cpp:260] Executor re-registered on agent 
e79deb05-5f20-48ca-9de9-a9610504e040-S1
{noformat}

Now i want to know how to remove these tasks, and where are these tasks 
running? (I do not think run in the mesos container because I specified 
`"type": "DOCKER"`)

There is a slave also thrown the following exception:

{noformat}
Message from syslogd@slave2 at May  6 08:27:35 ...
 kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1

Message from syslogd@slave2 at May  6 08:27:45 ...
 kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1
{noformat}

> Mesos marathon and docker not synchronized
> ------------------------------------------
>
>                 Key: MESOS-7466
>                 URL: https://issues.apache.org/jira/browse/MESOS-7466
>             Project: Mesos
>          Issue Type: Bug
>            Reporter: jasontom
>            Priority: Critical
>
> I submitted an application group in marathon, and then I deleted the 
> application group because there was some problem with the definition of the 
> application group, and then I submitted the application group again via the 
> marathon api. After multiple deletions and multiple submissions, I found some 
> problems (this is the third time I encountered the problem), I deleted the 
> application group in the marathon UI, found in the `deployments` there are 
> some content, and can not be deleted, suggesting that `error destroying null: 
> app '/ Null' does not exist `; in the mesos ui task status is `running`; 
> docker container status is `up`. Then, I restart mesos-master, mesos-slave, 
> marathon and zookeeper, the result is the same.
> I suggested that I clear the marathon in the zookeeper and then I did; 
> marathon was back to normal, just like reinstalling it. But the tasks I want 
> to delete are still in mesos ui and the state is `running`.
> After that, I cleared the meso in zookeeper, and the result is the same. The 
> task I want to delete is still in mesos ui. But I have to delete the task in 
> the docker does not exist, I do not know why.
> The follow is my config info:
> {noformat}
> $ curl -sSL http://172.30.30.4:5050/version | python -m json.tool
> {
>     "build_date": "2017-04-12 16:39:09",
>     "build_time": 1492015149.0,
>     "build_user": "centos",
>     "git_sha": "de306b5786de3c221bae1457c6f2ccaeb38eef9f",
>     "git_tag": "1.2.0",
>     "version": "1.2.0"
> }
> $ curl -sSL http://172.30.30.4:8080/v2/info | python -m json.tool
> {
>     ......
>     "name": "marathon",
>     "version": "1.4.2",
>     ......
> }
> $ docker version
> Client:
>  Version:      1.12.5
>  API version:  1.24
>  Go version:   go1.6.4
>  Git commit:   7392c3b
>  Built:        Fri Dec 16 02:23:59 2016
>  OS/Arch:      linux/amd64
> Server:
>  Version:      1.12.5
>  API version:  1.24
>  Go version:   go1.6.4
>  Git commit:   7392c3b
>  Built:        Fri Dec 16 02:23:59 2016
>  OS/Arch:      linux/amd64
> $ docker info
> Containers: 2
>  Running: 1
>  Paused: 0
>  Stopped: 1
> Images: 16
> Server Version: 1.12.5
> Storage Driver: devicemapper
>  Pool Name: docker-253:0-403269431-pool
>  Pool Blocksize: 65.54 kB
>  Base Device Size: 10.74 GB
>  Backing Filesystem: xfs
>  Data file: /dev/loop0
>  Metadata file: /dev/loop1
>  Data Space Used: 4.915 GB
>  Data Space Total: 107.4 GB
>  Data Space Available: 94.26 GB
>  Metadata Space Used: 7.115 MB
>  Metadata Space Total: 2.147 GB
>  Metadata Space Available: 2.14 GB
>  Thin Pool Minimum Free Space: 10.74 GB
>  Udev Sync Supported: true
>  Deferred Removal Enabled: false
>  Deferred Deletion Enabled: false
>  Deferred Deleted Device Count: 0
>  Data loop file: /var/lib/docker/devicemapper/devicemapper/data
>  WARNING: Usage of loopback devices is strongly discouraged for production 
> use. Use `--storage-opt dm.thinpooldev` to specify a custom block storage 
> device.
>  Metadata loop file: /var/lib/docker/devicemapper/devicemapper/metadata
>  Library Version: 1.02.135-RHEL7 (2016-09-28)
> Logging Driver: json-file
> Cgroup Driver: cgroupfs
> Plugins:
>  Volume: local
>  Network: calico null bridge host overlay
> Swarm: inactive
> Runtimes: runc
> Default Runtime: runc
> Security Options: seccomp
> Kernel Version: 3.10.0-514.2.2.el7.x86_64
> Operating System: CentOS Linux 7 (Core)
> OSType: linux
> Architecture: x86_64
> CPUs: 8
> Total Memory: 15.51 GiB
> Name: slave1
> ID: 4VRS:RPSC:ASAX:MTRJ:TOGC:6RFL:6RJS:J4MK:3VSN:JZO2:Q2FB:LJEW
> Docker Root Dir: /var/lib/docker
> Debug Mode (client): false
> Debug Mode (server): false
> Registry: https://index.docker.io/v1/
> Cluster Store: etcd://172.30.30.5:2379
> Cluster Advertise: 172.30.30.5:2375
> Insecure Registries:
>  172.30.30.8:80
>  127.0.0.0/8
> $ /usr/sbin/mesos-master \
>     --zk=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
>     --port=5050 \
>     --log_dir=/var/log/mesos \
>     --cluster=cp_cluster \
>     --hostname=172.30.30.4 \
>     --quorum=1 \
>     --work_dir=/var/lib/mesos
> $ /usr/sbin/mesos-slave \
>     --master=zk://172.30.30.4:2181,172.30.30.12:2181/mesos \
>     --log_dir=/var/log/mesos \
>     --containerizers=docker,mesos \
>     --executor_registration_timeout=5mins \
>     --hostname=172.30.30.5 \
>     --ip=172.30.30.5 \
>     --isolation=filesystem/linux,docker/runtime,network/cni \
>     --network_cni_config_dir=/var/lib/mesos/cni/config \
>     --network_cni_plugins_dir=/var/lib/mesos/cni/plugins \
>     --work_dir=/var/lib/mesos
> $ curl -sSL http://172.30.30.12:5050/master/tasks | python -m json.tool | 
> grep "\"id\""
>             "id": 
> "monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef",
>             "id": 
> "monitor-tools_prometheus.c6bdfccc-3233-11e7-bd32-024222b481ef",
>             "id": 
> "monitor-tools_alertmanager.c3bfff0b-3233-11e7-bd32-024222b481ef",
>             "id": 
> "monitor-exporter_node-exporter.0e44c1c5-3232-11e7-bd32-024222b481ef",
>             "id": 
> "monitor-exporter_cadvisor.0e455e07-3232-11e7-bd32-024222b481ef",
>             "id": 
> "monitor-exporter_mesos-exporter.0e43fe73-3232-11e7-bd32-024222b481ef",
>             "id": "syslog.e0645b43-3238-11e7-90c2-02423620ccdc",
>             "id": "syslog.e72bbd15-3238-11e7-90c2-02423620ccdc",
>             "id": "syslog.a0b162ff-316f-11e7-bd32-024222b481ef",
>             "id": "mesos-dns.1286663d-30ae-11e7-bd32-024222b481ef",
>             "id": "mesos-dns.12c20fae-30ae-11e7-bd32-024222b481ef",
>             "id": "mesos-dns.1286181c-30ae-11e7-bd32-024222b481ef",
> # slave1
> $ docker ps -a
> CONTAINER ID        IMAGE                        COMMAND             CREATED  
>            STATUS              PORTS               NAMES
> e05468eaedc2        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
> ago         Up 4 hours                              calico-node
> # slave2
> $ docker ps -a
> CONTAINER ID        IMAGE                        COMMAND             CREATED  
>            STATUS              PORTS               NAMES
> 7fcbb3f2fa61        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
> ago         Up 4 hours                              calico-node
> # slave3
> $ docker ps -a
> CONTAINER ID        IMAGE                        COMMAND             CREATED  
>            STATUS              PORTS               NAMES
> c69eba5bc322        quay.io/calico/node:v1.1.0   "start_runit"       2 weeks 
> ago         Up 4 hours                              calico-node
> {noformat}
> `monitor-***` tasks are I want to delete, but these tasks do not exist in the 
> docker (`docker ps -a` is not the task).
> The follow is `monitor-***`tasks log info(`stdout`):
> {noformat}
> ......
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Received killTask for task 
> monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
> Re-registered docker executor on 172.30.30.6
> Re-registered docker executor on 172.30.30.6
> {noformat}
> Now it's `stderr`:
> {noformat}
> I0506 23:26:45.211597 17482 exec.cpp:162] Version: 1.2.0
> I0506 23:26:45.224807 17489 exec.cpp:237] Executor registered on agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 23:26:45.310137 17486 docker.cpp:850] Running docker -H 
> unix:///var/run/docker.sock run --cpu-shares 512 --memory 1073741824 
> --env-file /tmp/qT2brz -v /etc/localtime:/etc/localtime:ro -v 
> /var/lib/mesos/slaves/e79deb05-5f20-48ca-9de9-a9610504e040-S1/frameworks/1e68ea0f-0f0b-4f14-8e2e-ab10169ee5f3-0000/executors/monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef/runs/7674214b-a609-47e6-a51a-7129616e8494:/mnt/mesos/sandbox
>  --net calico1 --log-driver=syslog 
> --log-opt=syslog-address=tcp://172.30.30.11:514 --log-opt=tag=grafana 
> --label=MESOS_TASK_ID=monitor-tools_grafana.ea8672ed-3233-11e7-bd32-024222b481ef
>  --name 
> mesos-e79deb05-5f20-48ca-9de9-a9610504e040-S1.7674214b-a609-47e6-a51a-7129616e8494
>  grafana/grafana
> I0506 16:57:13.687083 17489 exec.cpp:488] Agent exited, but framework has 
> checkpointing enabled. Waiting 15mins to reconnect with agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 16:57:14.359922 17486 exec.cpp:283] Received reconnect request from 
> agent e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 16:57:14.366843 17488 exec.cpp:260] Executor re-registered on agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 17:45:40.419311 17485 exec.cpp:488] Agent exited, but framework has 
> checkpointing enabled. Waiting 15mins to reconnect with agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 17:45:41.629977 17485 exec.cpp:283] Received reconnect request from 
> agent e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 17:45:41.658910 17485 exec.cpp:260] Executor re-registered on agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 18:02:20.707939 17485 exec.cpp:488] Agent exited, but framework has 
> checkpointing enabled. Waiting 15mins to reconnect with agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 18:03:43.669370 17485 exec.cpp:283] Received reconnect request from 
> agent e79deb05-5f20-48ca-9de9-a9610504e040-S1
> I0506 18:03:43.698617 17485 exec.cpp:260] Executor re-registered on agent 
> e79deb05-5f20-48ca-9de9-a9610504e040-S1
> {noformat}
> Now i want to know how to remove these tasks, and where are these tasks 
> running? (I do not think run in the mesos container because I specified 
> `"type": "DOCKER"`)
> There is a slave also thrown the following exception:
> {noformat}
> Message from syslogd@slave2 at May  6 08:27:35 ...
>  kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1
> Message from syslogd@slave2 at May  6 08:27:45 ...
>  kernel:unregister_netdevice: waiting for lo to become free. Usage count = 1
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to