Hello,

Would you please provide me with the executor's stderr log? This can be
found in the work directory on the agent, it should give us a bit more
information as to why it failed to start the task.

It'll be deeply nested, something like:

c:mesoswork_dirslaves7dc02270-a4e1-4f59-9ad7-56bad5182ea4-S3frameworkseb32cef4-c503-4ab7-85d4-8d4577e6a3bf-0000executorsnotepad.fcf078d1-084a-11e8-8f77-02421c3bc93crunslateststderr
(and stdout)

Thanks,

Andy

On 02/02/2018 1:30 pm, ajkf9uvxc ajkf9uvxc wrote: 

> Hi, 
> 
> I am trying to get a job in DCOS to run a docker container on a Windows agent 
> machine. DCOS was installed using the AWS CF template here: 
> https://downloads.dcos.io/dcos/stable/aws.html [1] (single master). 
> 
> The Windows agent is added: 
> 
> C:mesosmesosbuildsrcmesos-agent.exe --attributes=os:windows 
> --containerizers=docker,mesos --hostname=10.19.10.206 --IP=10.19.10.206 
> --master=zk://10.22.1.94:2181/mesos [2] --work_dir=c:mesoswork_dir 
> --launcher_dir=c:mesosmesosbuildsrc --log_dir=c:mesoslogs 
> 
> And a simple job works: 
> 
> dcos.activestate.com [3] -> Job -> New 
> 
> { 
> 
> "id": "mywindowstest01", 
> 
> "labels": {}, 
> 
> "run": { 
> 
> "cpus": 0.01, 
> 
> "mem": 128, 
> 
> "disk": 0, 
> 
> "cmd": "C:\Windows\System32\cmd.exe /c echo helloworld > 
> c:\mesos\work_dir\helloworld2", 
> 
> "env": {}, 
> 
> "placement": { 
> 
> "constraints": [ 
> 
> { 
> 
> "attribute": "os", 
> 
> "operator": "EQ", 
> 
> "value": "windows" 
> 
> } 
> 
> ] 
> 
> }, 
> 
> "artifacts": [], 
> 
> "maxLaunchDelay": 3600, 
> 
> "volumes": [], 
> 
> "restart": { 
> 
> "policy": "NEVER" 
> 
> } 
> 
> }, 
> 
> "schedules": [] 
> 
> } 
> 
> creates: "c:\mesos\work_dir\helloworld2" 
> 
> The Windows agent has DockerCE installed and is set to run Windows containers 
> (tried with Linux containers as well and getting the same problem, but for 
> the purpose of this question let's stick to Windows containers) 
> 
> I confirmed that it's possible to run a Windows container manually, directly 
> on Windows 10 by starting a Powershell as Administrator and running: 
> 
> docker run -ti microsoft/windowsservercore 
> and 
> 
> docker run microsoft/windowsservercore 
> 
> Both commands create a new container (verified with "docker ps" , besides I 
> get a cmd.exe shell in the conatiner for the first command) 
> 
> Now the problem: 
> 
> trying to run a container from DCOS does not work: 
> 
> dcos job add a.json 
> 
> with the json: 
> 
> { 
> "id": "myattempt11", 
> "labels": {}, 
> "run": { 
> "env": {}, 
> "cpus": 1.00, 
> "mem": 512, 
> "disk": 1000, 
> "placement": { 
> "constraints": [ 
> { 
> "attribute": "os", 
> "operator": "EQ", 
> "value": "windows" 
> } 
> ] 
> }, 
> "artifacts": [], 
> "maxLaunchDelay": 3600, 
> "docker": { 
> "image": "microsoft/windowsservercore" 
> }, 
> "restart": { 
> "policy": "NEVER" 
> } 
> }, 
> "schedules": [] 
> } 
> 
> Does not work: 
> 
> # dcos job add a.json 
> 
> # dcos job run myattempt11 
> Run ID: 20180202203339zVpxc 
> 
> The log on the Mesos Agent on Windows shows activity but not much information 
> about the problem (see "TASK_FAILED" at the end below): 
> 
> Log file created at: 2018/02/02 12:52:47 
> Running on machine: DESKTOP-JJK06UJ 
> Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg 
> I0202 12:52:47.330880 8388 logging.cpp:201] INFO level logging started! 
> I0202 12:52:47.335886 8388 main.cpp:365] Build: 2017-12-20 23:35:42 UTC by 
> Anne S Bell 
> I0202 12:52:47.335886 8388 main.cpp:366] Version: 1.5.0 
> I0202 12:52:47.337895 8388 main.cpp:373] Git SHA: 
> 327726d3c7272806c8f3c3b7479758c26e55fd43 
> I0202 12:52:47.358888 8388 resolver.cpp:69] Creating default secret resolver 
> I0202 12:52:47.574883 8388 containerizer.cpp:304] Using isolation { 
> windows/cpu, filesystem/windows, windows/mem, environment_secret } 
> I0202 12:52:47.577883 8388 provisioner.cpp:299] Using default backend 'copy' 
> I0202 12:52:47.596886 3348 slave.cpp:262] Mesos agent started on 
> (1)@10.19.10.206:5051 [4] 
> I0202 12:52:47.597883 3348 slave.cpp:263] Flags at startup: 
> --appc_simple_discovery_uri_prefix="http://"; 
> --appc_store_dir="C:UsersactiveitAppDataLocalTempmesosstoreappc" 
> --attributes="os:windows" --authenticate_http_readonly="false" 
> --authenticate_http_readwrite="false" --authenticatee="crammd5" 
> --authentication_backoff_factor="1secs" --authorizer="local" 
> --container_disk_watch_interval="15secs" --containerizers="docker,mesos" 
> --default_role="*" --disk_watch_interval="1mins" --docker="docker" 
> --docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io 
> [5]" --docker_remove_delay="6hrs" --docker_socket="//./pipe/docker_engine" 
> --docker_stop_timeout="0ns" 
> --docker_store_dir="C:UsersactiveitAppDataLocalTempmesosstoredocker" 
> --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
> --enforce_container_disk_quota="false" 
> --executor_registration_timeout="1mins" 
> --executor_reregistration_timeout="2secs" 
> --executor_shutdown_grace_period="5secs"
--fetcher_cache_dir="C:UsersactiveitAppDataLocalTempmesosfetch" 
--fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname="10.19.10.206" --hostname_lookup="true" 
--http_command_executor="false" --http_heartbeat_interval="30secs" 
--initialize_driver_logging="true" --ip="10.19.10.206" 
--isolation="windows/cpu,windows/mem" --launcher="windows" 
--launcher_dir="c:mesosmesosbuildsrc" --log_dir="c:mesoslogs" --logbufsecs="0" 
--logging_level="INFO" --master="zk://10.22.1.94:2181/mesos [2]" 
--max_completed_executors_per_framework="150" 
--oversubscribed_resources_interval="15secs" --port="5051" 
--qos_correction_interval_min="0ns" --quiet="false" 
--reconfiguration_policy="equal" --recover="reconnect" 
--recovery_timeout="15mins" --registration_backoff_factor="1secs" 
--runtime_dir="C:ProgramDatamesosruntime" --sandbox_directory="C:mesossandbox" 
--strict="true" --version="false" --work_dir="c:mesoswork_dir"
--zk_session_timeout="10secs" 
> I0202 12:52:47.604887 3348 slave.cpp:612] Agent resources: 
> [{"name":"cpus","scalar":{"value":4.0},"type":"SCALAR"},{"name":"mem","scalar":{"value":15290.0},"type":"SCALAR"},{"name":"disk","scalar":{"value":470301.0},"type":"SCALAR"},{"name":"ports","ranges":{"range":[{"begin":31000,"end":32000}]},"type":"RANGES"}]
>  
> I0202 12:52:47.725885 3348 slave.cpp:620] Agent attributes: [ os=windows ] 
> I0202 12:52:47.727886 3348 slave.cpp:629] Agent hostname: 10.19.10.206 
> I0202 12:52:47.735886 7652 task_status_update_manager.cpp:181] Pausing 
> sending task status updates 
> I0202 12:52:47.738890 4052 group.cpp:341] Group process 
> (zookeeper-group(1)@10.19.10.206:5051 [4]) connected to ZooKeeper 
> I0202 12:52:47.739887 4052 group.cpp:831] Syncing group operations: queue 
> size (joins, cancels, datas) = (0, 0, 0) 
> I0202 12:52:47.740885 4052 group.cpp:419] Trying to create path '/mesos' in 
> ZooKeeper 
> I0202 12:52:47.773885 5168 state.cpp:66] Recovering state from 
> 'c:mesoswork_dirmeta' 
> E0202 12:52:47.773885 3348 slave.cpp:1009] Failed to attach 
> 'c:mesoslogsmesos-agent.exe.INFO [6]' to virtual path '/slave/log': Failed to 
> get realpath of 'c:mesoslogsmesos-agent.exe.INFO [6]': Failed to get 
> attributes for file 'c:mesoslogsmesos-agent.exe.INFO [6]': The system cannot 
> find the file specified. 
> 
> I0202 12:52:47.774884 5168 state.cpp:724] No committed checkpointed resources 
> found at 'c:mesoswork_dirmetaresourcesresources.info [7]' 
> I0202 12:52:47.779883 5168 state.cpp:110] Failed to find the latest agent 
> from 'c:mesoswork_dirmeta' 
> I0202 12:52:47.781888 3528 task_status_update_manager.cpp:207] Recovering 
> task status update manager 
> I0202 12:52:47.782883 3348 docker.cpp:890] Recovering Docker containers 
> I0202 12:52:47.782883 7652 containerizer.cpp:674] Recovering containerizer 
> I0202 12:52:47.807888 3768 provisioner.cpp:495] Provisioner recovery complete 
> I0202 12:52:47.891667 5168 detector.cpp:152] Detected a new leader: 
> (id='1171') 
> I0202 12:52:47.892666 7652 group.cpp:700] Trying to get 
> '/mesos/json.info_0000001171' in ZooKeeper 
> I0202 12:52:47.970657 5168 zookeeper.cpp:262] A new leading master 
> (UPID=master@10.22.1.94:5050 [8]) is detected 
> I0202 12:52:48.011252 7652 slave.cpp:6776] Finished recovery 
> I0202 12:52:48.020246 3768 task_status_update_manager.cpp:181] Pausing 
> sending task status updates 
> I0202 12:52:48.020246 7652 slave.cpp:1055] New master detected at 
> master@10.22.1.94:5050 [8] 
> I0202 12:52:48.021251 7652 slave.cpp:1099] No credentials provided. 
> Attempting to register without authentication 
> I0202 12:52:48.023254 7652 slave.cpp:1110] Detecting new master 
> I0202 12:52:48.330085 4052 slave.cpp:1275] Registered with master 
> master@10.22.1.94:5050 [8]; given agent ID 
> a0664e60-846a-42d0-9586-cf97e997eba3-S0 
> I0202 12:52:48.331082 5168 task_status_update_manager.cpp:188] Resuming 
> sending task status updates 
> I0202 12:52:48.348086 4052 slave.cpp:1352] Forwarding agent update 
> {"offer_operations":{},"resource_version_uuid":{"value":"DEVEk/KOR5KLtmOgVG9qvw=="},"slave_id":{"value":"a0664e60-846a-42d0-9586-cf97e997eba3-S0"},"update_oversubscribed_resources":true}
>  
> W0202 12:52:48.351085 4052 slave.cpp:1334] Already registered with master 
> master@10.22.1.94:5050 [8] 
> I0202 12:52:48.356086 4052 slave.cpp:1352] Forwarding agent update 
> {"offer_operations":{},"resource_version_uuid":{"value":"DEVEk/KOR5KLtmOgVG9qvw=="},"slave_id":{"value":"a0664e60-846a-42d0-9586-cf97e997eba3-S0"},"update_oversubscribed_resources":true}
>  
> W0202 12:52:48.358086 4052 slave.cpp:1334] Already registered with master 
> master@10.22.1.94:5050 [8] 
> I0202 12:52:48.359086 4052 slave.cpp:1352] Forwarding agent update 
> {"offer_operations":{},"resource_version_uuid":{"value":"DEVEk/KOR5KLtmOgVG9qvw=="},"slave_id":{"value":"a0664e60-846a-42d0-9586-cf97e997eba3-S0"},"update_oversubscribed_resources":true}
>  
> W0202 12:52:48.362089 4052 slave.cpp:1334] Already registered with master 
> master@10.22.1.94:5050 [8] 
> I0202 12:52:48.363085 4052 slave.cpp:1352] Forwarding agent update 
> {"offer_operations":{},"resource_version_uuid":{"value":"DEVEk/KOR5KLtmOgVG9qvw=="},"slave_id":{"value":"a0664e60-846a-42d0-9586-cf97e997eba3-S0"},"update_oversubscribed_resources":true}
>  
> W0202 12:52:48.364082 4052 slave.cpp:1334] Already registered with master 
> master@10.22.1.94:5050 [8] 
> I0202 12:52:48.365085 4052 slave.cpp:1352] Forwarding agent update 
> {"offer_operations":{},"resource_version_uuid":{"value":"DEVEk/KOR5KLtmOgVG9qvw=="},"slave_id":{"value":"a0664e60-846a-42d0-9586-cf97e997eba3-S0"},"update_oversubscribed_resources":true}
>  
> I0202 12:52:50.938498 7652 slave.cpp:1831] Got assigned task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' for 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:50.962504 7652 slave.cpp:2101] Authorizing task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' for 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:50.965504 3768 slave.cpp:2494] Launching task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' for 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:50.988512 3768 slave.cpp:8373] Launching executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 with resources 
> [{"allocation_info":{"role":"*"},"name":"cpus","scalar":{"value":0.1},"type":"SCALAR"},{"allocation_info":{"role":"*"},"name":"mem","scalar":{"value":32.0},"type":"SCALAR"}]
>  in work directory 
> 'c:mesoswork_dirslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88runs74298e92-9700-486d-b211-a42e5fd0bf85'
>  
> I0202 12:52:50.995501 3768 slave.cpp:3046] Launching container 
> 74298e92-9700-486d-b211-a42e5fd0bf85 for executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:51.010500 3768 slave.cpp:2580] Queued task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' for 
> executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:51.017498 3348 docker.cpp:1144] Starting container 
> '74298e92-9700-486d-b211-a42e5fd0bf85' for task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' (and 
> executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88') of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:53.731667 1104 docker.cpp:784] Checkpointing pid 7732 to 
> 'c:mesoswork_dirmetaslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88runs74298e92-9700-486d-b211-a42e5fd0bf85pidsforked.pid'
>  
> I0202 12:52:53.894371 4052 slave.cpp:4314] Got registration for executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 from 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:53.911371 1104 docker.cpp:1627] Ignoring updating container 
> 74298e92-9700-486d-b211-a42e5fd0bf85 because resources passed to update are 
> identical to existing resources 
> I0202 12:52:53.914371 3768 slave.cpp:2785] Sending queued task 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' to 
> executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 at 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:53.931371 7652 slave.cpp:4771] Handling status update 
> TASK_STARTING (Status UUID: ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 from 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:53.942371 5168 task_status_update_manager.cpp:328] Received task 
> status update TASK_STARTING (Status UUID: 
> ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:53.948371 5168 task_status_update_manager.cpp:842] Checkpointing 
> UPDATE for task status update TASK_STARTING (Status UUID: 
> ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:53.950371 1104 slave.cpp:5254] Forwarding the update 
> TASK_STARTING (Status UUID: ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 to master@10.22.1.94:5050 
> [8] 
> I0202 12:52:53.953371 1104 slave.cpp:5163] Sending acknowledgement for status 
> update TASK_STARTING (Status UUID: ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for 
> task myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 to 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:54.049816 3348 task_status_update_manager.cpp:401] Received task 
> status update acknowledgement (UUID: ef5adc2f-6f66-44c3-bc98-7697c1315ebf) 
> for task myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 
> of framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:54.051817 3348 task_status_update_manager.cpp:842] Checkpointing 
> ACK for task status update TASK_STARTING (Status UUID: 
> ef5adc2f-6f66-44c3-bc98-7697c1315ebf) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:59.255755 4052 slave.cpp:4771] Handling status update TASK_FAILED 
> (Status UUID: c0775c86-4f1b-44a6-ae8f-347486f6fa9f) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 from 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:59.260759 4052 task_status_update_manager.cpp:328] Received task 
> status update TASK_FAILED (Status UUID: c0775c86-4f1b-44a6-ae8f-347486f6fa9f) 
> for task myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 
> of framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:59.261757 4052 task_status_update_manager.cpp:842] Checkpointing 
> UPDATE for task status update TASK_FAILED (Status UUID: 
> c0775c86-4f1b-44a6-ae8f-347486f6fa9f) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:59.263756 5168 slave.cpp:5254] Forwarding the update TASK_FAILED 
> (Status UUID: c0775c86-4f1b-44a6-ae8f-347486f6fa9f) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 to master@10.22.1.94:5050 
> [8] 
> I0202 12:52:59.265756 5168 slave.cpp:5163] Sending acknowledgement for status 
> update TASK_FAILED (Status UUID: c0775c86-4f1b-44a6-ae8f-347486f6fa9f) for 
> task myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 to 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:52:59.367189 7052 task_status_update_manager.cpp:401] Received task 
> status update acknowledgement (UUID: c0775c86-4f1b-44a6-ae8f-347486f6fa9f) 
> for task myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 
> of framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:52:59.368187 7052 task_status_update_manager.cpp:842] Checkpointing 
> ACK for task status update TASK_FAILED (Status UUID: 
> c0775c86-4f1b-44a6-ae8f-347486f6fa9f) for task 
> myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88 of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:53:00.261153 4052 slave.cpp:5386] Got exited event for 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:53:00.471400 7052 docker.cpp:2415] Executor for container 
> 74298e92-9700-486d-b211-a42e5fd0bf85 has exited 
> I0202 12:53:00.472362 7052 docker.cpp:2186] Destroying container 
> 74298e92-9700-486d-b211-a42e5fd0bf85 in RUNNING state 
> I0202 12:53:00.474362 7052 docker.cpp:2236] Running docker stop on container 
> 74298e92-9700-486d-b211-a42e5fd0bf85 
> I0202 12:53:00.477478 3348 slave.cpp:5795] Executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 exited with status 0 
> I0202 12:53:00.478476 3348 slave.cpp:5899] Cleaning up executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' of 
> framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 at 
> executor(1)@10.19.10.206:49855 [9] 
> I0202 12:53:00.481472 4052 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88runs74298e92-9700-486d-b211-a42e5fd0bf85'
>  for gc 6.99989026072889days in the future 
> I0202 12:53:00.483475 3528 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88'
>  for gc 6.99987866347259days in the future 
> I0202 12:53:00.484474 5168 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirmetaslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88runs74298e92-9700-486d-b211-a42e5fd0bf85'
>  for gc 6.99999439265185days in the future 
> I0202 12:53:00.485474 5168 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirmetaslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000executorsmyattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88'
>  for gc 6.99987864033482days in the future 
> I0202 12:53:00.485474 3348 slave.cpp:6006] Cleaning up framework 
> 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:53:00.486479 1104 task_status_update_manager.cpp:289] Closing task 
> status update streams for framework 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 
> I0202 12:53:00.487473 3768 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000'
>  for gc 6.9998786172days in the future 
> I0202 12:53:00.488477 3768 gc.cpp:90] Scheduling 
> 'c:mesoswork_dirmetaslavesa0664e60-846a-42d0-9586-cf97e997eba3-S0frameworksca2eae6-8912-4f6a-984a-d501ac02ff88-0000'
>  for gc 6.99987860557926days in the future 
> I0202 12:53:47.742332 7052 slave.cpp:6314] Current disk usage 24.73%. Max 
> allowed age: 4.568714599279827days 
> I0202 12:54:01.675030 7052 slave.cpp:6222] Framework 
> 0ca2eae6-8912-4f6a-984a-d501ac02ff88-0000 seems to have exited. Ignoring 
> registration timeout for executor 
> 'myattempt11_20180202203339zVpxc.07298e1c-085b-11e8-bc6d-ae95ed0c8d88' 
> I0202 12:54:03.169529 3348 slave.cpp:970] Received SIGUSR1 signal; 
> unregistering and shutting down 
> I0202 12:54:03.170536 3348 slave.cpp:931] Agent terminating 
> I0202 12:54:03.199530 3308 process.cpp:887] Failed to accept socket: future 
> discarded 
> 
> in DCOS web-ui -> Jobs -> myattempt11 -> Run History there is also no 
> information. 
> 
> Are there any good troubleshooting tips / ideas what to try or where to find 
> more informative logs to run a Docker container on Windows using Mesos? 
> 
> Are there any more suitable alternative orchestration tools to run Docker 
> Windows containers in a cluster?
 

Links:
------
[1] https://downloads.dcos.io/dcos/stable/aws.html
[2] http://10.22.1.94:2181/mesos
[3] http://dcos.activestate.com/
[4] http://10.19.10.206:5051/
[5] https://registry-1.docker.io/
[6] http://mesos-agent.exe.info/
[7] http://resources.info/
[8] http://master@10.22.1.94:5050/
[9] http://10.19.10.206:49855/

Reply via email to