[
https://issues.apache.org/jira/browse/MESOS-4832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jie Yu updated MESOS-4832:
--------------------------
Fix Version/s: 0.28.0
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes exits
> when the /tmp directory is bind-mounted
> ------------------------------------------------------------------------------------------------------------------
>
> Key: MESOS-4832
> URL: https://issues.apache.org/jira/browse/MESOS-4832
> Project: Mesos
> Issue Type: Bug
> Components: containerization, docker
> Affects Versions: 0.27.0
> Environment: Seen on CentOS 7 & Debian 8.
> Reporter: Joseph Wu
> Assignee: Jie Yu
> Labels: mesosphere, test
> Fix For: 0.28.0
>
>
> If the {{/tmp}} directory (where Mesos tests create temporary directories) is
> a bind mount, the test suite will exit here:
> {code}
> [ RUN ]
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes
> I0226 03:17:26.722806 1097 leveldb.cpp:174] Opened db in 12.587676ms
> I0226 03:17:26.723496 1097 leveldb.cpp:181] Compacted db in 636999ns
> I0226 03:17:26.723536 1097 leveldb.cpp:196] Created db iterator in 18271ns
> I0226 03:17:26.723547 1097 leveldb.cpp:202] Seeked to beginning of db in
> 1555ns
> I0226 03:17:26.723554 1097 leveldb.cpp:271] Iterated through 0 keys in the
> db in 363ns
> I0226 03:17:26.723593 1097 replica.cpp:779] Replica recovered with log
> positions 0 -> 0 with 1 holes and 0 unlearned
> I0226 03:17:26.724128 1117 recover.cpp:447] Starting replica recovery
> I0226 03:17:26.724367 1117 recover.cpp:473] Replica is in EMPTY status
> I0226 03:17:26.725237 1117 replica.cpp:673] Replica in EMPTY status received
> a broadcasted recover request from (13810)@172.30.2.151:51934
> I0226 03:17:26.725744 1114 recover.cpp:193] Received a recover response from
> a replica in EMPTY status
> I0226 03:17:26.726356 1111 master.cpp:376] Master
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a (ip-172-30-2-151.mesosphere.io) started
> on 172.30.2.151:51934
> I0226 03:17:26.726369 1118 recover.cpp:564] Updating replica status to
> STARTING
> I0226 03:17:26.726378 1111 master.cpp:378] Flags at startup: --acls=""
> --allocation_interval="1secs" --allocator="HierarchicalDRF"
> --authenticate="true" --authenticate_http="true" --authenticate_slaves="true"
> --authenticators="crammd5" --authorizers="local"
> --credentials="/tmp/djHTVQ/credentials" --framework_sorter="drf"
> --help="false" --hostname_lookup="true" --http_authenticators="basic"
> --initialize_driver_logging="true" --log_auto_initialize="true"
> --logbufsecs="0" --logging_level="INFO" --max_completed_frameworks="50"
> --max_completed_tasks_per_framework="1000" --max_slave_ping_timeouts="5"
> --quiet="false" --recovery_slave_removal_limit="100%"
> --registry="replicated_log" --registry_fetch_timeout="1mins"
> --registry_store_timeout="100secs" --registry_strict="true"
> --root_submissions="true" --slave_ping_timeout="15secs"
> --slave_reregister_timeout="10mins" --user_sorter="drf" --version="false"
> --webui_dir="/usr/local/share/mesos/webui" --work_dir="/tmp/djHTVQ/master"
> --zk_session_timeout="10secs"
> I0226 03:17:26.726605 1111 master.cpp:423] Master only allowing
> authenticated frameworks to register
> I0226 03:17:26.726616 1111 master.cpp:428] Master only allowing
> authenticated slaves to register
> I0226 03:17:26.726632 1111 credentials.hpp:35] Loading credentials for
> authentication from '/tmp/djHTVQ/credentials'
> I0226 03:17:26.726860 1111 master.cpp:468] Using default 'crammd5'
> authenticator
> I0226 03:17:26.726977 1111 master.cpp:537] Using default 'basic' HTTP
> authenticator
> I0226 03:17:26.727092 1111 master.cpp:571] Authorization enabled
> I0226 03:17:26.727243 1118 hierarchical.cpp:144] Initialized hierarchical
> allocator process
> I0226 03:17:26.727285 1116 whitelist_watcher.cpp:77] No whitelist given
> I0226 03:17:26.728852 1114 master.cpp:1712] The newly elected leader is
> [email protected]:51934 with id 5cc57c0e-f1ad-4107-893f-420ed1a1db1a
> I0226 03:17:26.728876 1114 master.cpp:1725] Elected as the leading master!
> I0226 03:17:26.728891 1114 master.cpp:1470] Recovering from registrar
> I0226 03:17:26.728977 1117 registrar.cpp:307] Recovering registrar
> I0226 03:17:26.731503 1112 leveldb.cpp:304] Persisting metadata (8 bytes) to
> leveldb took 4.977811ms
> I0226 03:17:26.731539 1112 replica.cpp:320] Persisted replica status to
> STARTING
> I0226 03:17:26.731711 1111 recover.cpp:473] Replica is in STARTING status
> I0226 03:17:26.732501 1114 replica.cpp:673] Replica in STARTING status
> received a broadcasted recover request from (13812)@172.30.2.151:51934
> I0226 03:17:26.732862 1111 recover.cpp:193] Received a recover response from
> a replica in STARTING status
> I0226 03:17:26.733264 1117 recover.cpp:564] Updating replica status to VOTING
> I0226 03:17:26.733836 1118 leveldb.cpp:304] Persisting metadata (8 bytes) to
> leveldb took 388246ns
> I0226 03:17:26.733855 1118 replica.cpp:320] Persisted replica status to
> VOTING
> I0226 03:17:26.733979 1113 recover.cpp:578] Successfully joined the Paxos
> group
> I0226 03:17:26.734149 1113 recover.cpp:462] Recover process terminated
> I0226 03:17:26.734478 1111 log.cpp:659] Attempting to start the writer
> I0226 03:17:26.735523 1114 replica.cpp:493] Replica received implicit
> promise request from (13813)@172.30.2.151:51934 with proposal 1
> I0226 03:17:26.736130 1114 leveldb.cpp:304] Persisting metadata (8 bytes) to
> leveldb took 576451ns
> I0226 03:17:26.736150 1114 replica.cpp:342] Persisted promised to 1
> I0226 03:17:26.736709 1115 coordinator.cpp:238] Coordinator attempting to
> fill missing positions
> I0226 03:17:26.737771 1114 replica.cpp:388] Replica received explicit
> promise request from (13814)@172.30.2.151:51934 for position 0 with proposal 2
> I0226 03:17:26.738386 1114 leveldb.cpp:341] Persisting action (8 bytes) to
> leveldb took 583184ns
> I0226 03:17:26.738404 1114 replica.cpp:712] Persisted action at 0
> I0226 03:17:26.739312 1118 replica.cpp:537] Replica received write request
> for position 0 from (13815)@172.30.2.151:51934
> I0226 03:17:26.739367 1118 leveldb.cpp:436] Reading position from leveldb
> took 26157ns
> I0226 03:17:26.740638 1118 leveldb.cpp:341] Persisting action (14 bytes) to
> leveldb took 1.238477ms
> I0226 03:17:26.740669 1118 replica.cpp:712] Persisted action at 0
> I0226 03:17:26.741158 1118 replica.cpp:691] Replica received learned notice
> for position 0 from @0.0.0.0:0
> I0226 03:17:26.742878 1118 leveldb.cpp:341] Persisting action (16 bytes) to
> leveldb took 1.697254ms
> I0226 03:17:26.742902 1118 replica.cpp:712] Persisted action at 0
> I0226 03:17:26.742916 1118 replica.cpp:697] Replica learned NOP action at
> position 0
> I0226 03:17:26.743393 1117 log.cpp:675] Writer started with ending position 0
> I0226 03:17:26.744370 1112 leveldb.cpp:436] Reading position from leveldb
> took 34329ns
> I0226 03:17:26.745240 1117 registrar.cpp:340] Successfully fetched the
> registry (0B) in 16.21888ms
> I0226 03:17:26.745350 1117 registrar.cpp:439] Applied 1 operations in
> 30460ns; attempting to update the 'registry'
> I0226 03:17:26.746016 1111 log.cpp:683] Attempting to append 210 bytes to
> the log
> I0226 03:17:26.746119 1116 coordinator.cpp:348] Coordinator attempting to
> write APPEND action at position 1
> I0226 03:17:26.746798 1114 replica.cpp:537] Replica received write request
> for position 1 from (13816)@172.30.2.151:51934
> I0226 03:17:26.747251 1114 leveldb.cpp:341] Persisting action (229 bytes) to
> leveldb took 411333ns
> I0226 03:17:26.747269 1114 replica.cpp:712] Persisted action at 1
> I0226 03:17:26.747808 1113 replica.cpp:691] Replica received learned notice
> for position 1 from @0.0.0.0:0
> I0226 03:17:26.749511 1113 leveldb.cpp:341] Persisting action (231 bytes) to
> leveldb took 1.673488ms
> I0226 03:17:26.749534 1113 replica.cpp:712] Persisted action at 1
> I0226 03:17:26.749550 1113 replica.cpp:697] Replica learned APPEND action at
> position 1
> I0226 03:17:26.750422 1111 registrar.cpp:484] Successfully updated the
> 'registry' in 5.021952ms
> I0226 03:17:26.750560 1111 registrar.cpp:370] Successfully recovered
> registrar
> I0226 03:17:26.750635 1112 log.cpp:702] Attempting to truncate the log to 1
> I0226 03:17:26.750751 1113 coordinator.cpp:348] Coordinator attempting to
> write TRUNCATE action at position 2
> I0226 03:17:26.751096 1116 master.cpp:1522] Recovered 0 slaves from the
> Registry (171B) ; allowing 10mins for slaves to re-register
> I0226 03:17:26.751126 1111 hierarchical.cpp:171] Skipping recovery of
> hierarchical allocator: nothing to recover
> I0226 03:17:26.751561 1118 replica.cpp:537] Replica received write request
> for position 2 from (13817)@172.30.2.151:51934
> I0226 03:17:26.751999 1118 leveldb.cpp:341] Persisting action (16 bytes) to
> leveldb took 406823ns
> I0226 03:17:26.752018 1118 replica.cpp:712] Persisted action at 2
> I0226 03:17:26.752521 1113 replica.cpp:691] Replica received learned notice
> for position 2 from @0.0.0.0:0
> I0226 03:17:26.754161 1113 leveldb.cpp:341] Persisting action (18 bytes) to
> leveldb took 1.614888ms
> I0226 03:17:26.754210 1113 leveldb.cpp:399] Deleting ~1 keys from leveldb
> took 26384ns
> I0226 03:17:26.754225 1113 replica.cpp:712] Persisted action at 2
> I0226 03:17:26.754240 1113 replica.cpp:697] Replica learned TRUNCATE action
> at position 2
> I0226 03:17:26.765103 1115 slave.cpp:193] Slave started on
> 399)@172.30.2.151:51934
> I0226 03:17:26.765130 1115 slave.cpp:194] Flags at startup:
> --appc_simple_discovery_uri_prefix="http://"
> --appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5"
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false"
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false"
> --cgroups_root="mesos" --container_disk_watch_interval="15secs"
> --containerizers="mesos"
> --credential="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential"
> --default_role="*" --disk_watch_interval="1mins" --docker="docker"
> --docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io"
> --docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock"
> --docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker"
> --enforce_container_disk_quota="false"
> --executor_registration_timeout="1mins"
> --executor_shutdown_grace_period="5secs"
> --fetcher_cache_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/fetch"
> --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks"
> --gc_disk_headroom="0.1" --hadoop_home="" --help="false"
> --hostname_lookup="true" --image_provisioner_backend="copy"
> --initialize_driver_logging="true" --isolation="posix/cpu,posix/mem"
> --launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src"
> --logbufsecs="0" --logging_level="INFO"
> --oversubscribed_resources_interval="15secs" --perf_duration="10secs"
> --perf_interval="1mins" --qos_correction_interval_min="0ns" --quiet="false"
> --recover="reconnect" --recovery_timeout="15mins"
> --registration_backoff_factor="10ms"
> --resources="cpu:2;mem:2048;disk(role1):2048"
> --revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox"
> --strict="true" --switch_user="true" --systemd_enable_support="true"
> --systemd_runtime_directory="/run/systemd/system" --version="false"
> --work_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP"
> I0226 03:17:26.765403 1115 credentials.hpp:83] Loading credential for
> authentication from
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential'
> I0226 03:17:26.765573 1115 slave.cpp:324] Slave using credential for:
> test-principal
> I0226 03:17:26.765733 1115 resources.cpp:576] Parsing resources as JSON
> failed: cpu:2;mem:2048;disk(role1):2048
> Trying semicolon-delimited string format instead
> I0226 03:17:26.766185 1115 slave.cpp:464] Slave resources: cpu(*):2;
> mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000]
> I0226 03:17:26.766242 1115 slave.cpp:472] Slave attributes: [ ]
> I0226 03:17:26.766250 1115 slave.cpp:477] Slave hostname:
> ip-172-30-2-151.mesosphere.io
> I0226 03:17:26.767325 1097 sched.cpp:222] Version: 0.28.0
> I0226 03:17:26.767390 1111 state.cpp:58] Recovering state from
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta'
> I0226 03:17:26.767603 1115 status_update_manager.cpp:200] Recovering status
> update manager
> I0226 03:17:26.767865 1113 docker.cpp:726] Recovering Docker containers
> I0226 03:17:26.767971 1111 sched.cpp:326] New master detected at
> [email protected]:51934
> I0226 03:17:26.768045 1111 sched.cpp:382] Authenticating with master
> [email protected]:51934
> I0226 03:17:26.768059 1111 sched.cpp:389] Using default CRAM-MD5
> authenticatee
> I0226 03:17:26.768070 1118 slave.cpp:4565] Finished recovery
> I0226 03:17:26.768273 1112 authenticatee.cpp:121] Creating new client SASL
> connection
> I0226 03:17:26.768435 1118 slave.cpp:4737] Querying resource estimator for
> oversubscribable resources
> I0226 03:17:26.768565 1111 master.cpp:5526] Authenticating
> [email protected]:51934
> I0226 03:17:26.768661 1118 slave.cpp:796] New master detected at
> [email protected]:51934
> I0226 03:17:26.768659 1115 authenticator.cpp:413] Starting authentication
> session for crammd5_authenticatee(839)@172.30.2.151:51934
> I0226 03:17:26.768679 1113 status_update_manager.cpp:174] Pausing sending
> status updates
> I0226 03:17:26.768728 1118 slave.cpp:859] Authenticating with master
> [email protected]:51934
> I0226 03:17:26.768743 1118 slave.cpp:864] Using default CRAM-MD5
> authenticatee
> I0226 03:17:26.768865 1118 slave.cpp:832] Detecting new master
> I0226 03:17:26.768868 1112 authenticator.cpp:98] Creating new server SASL
> connection
> I0226 03:17:26.768908 1114 authenticatee.cpp:121] Creating new client SASL
> connection
> I0226 03:17:26.769003 1118 slave.cpp:4751] Received oversubscribable
> resources from the resource estimator
> I0226 03:17:26.769103 1115 authenticatee.cpp:212] Received SASL
> authentication mechanisms: CRAM-MD5
> I0226 03:17:26.769131 1115 authenticatee.cpp:238] Attempting to authenticate
> with mechanism 'CRAM-MD5'
> I0226 03:17:26.769209 1116 master.cpp:5526] Authenticating
> slave(399)@172.30.2.151:51934
> I0226 03:17:26.769253 1114 authenticator.cpp:203] Received SASL
> authentication start
> I0226 03:17:26.769295 1115 authenticator.cpp:413] Starting authentication
> session for crammd5_authenticatee(840)@172.30.2.151:51934
> I0226 03:17:26.769307 1114 authenticator.cpp:325] Authentication requires
> more steps
> I0226 03:17:26.769403 1117 authenticatee.cpp:258] Received SASL
> authentication step
> I0226 03:17:26.769495 1114 authenticator.cpp:98] Creating new server SASL
> connection
> I0226 03:17:26.769531 1115 authenticator.cpp:231] Received SASL
> authentication step
> I0226 03:17:26.769554 1115 auxprop.cpp:107] Request to lookup properties for
> user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN:
> 'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false
> I0226 03:17:26.769562 1115 auxprop.cpp:179] Looking up auxiliary property
> '*userPassword'
> I0226 03:17:26.769608 1115 auxprop.cpp:179] Looking up auxiliary property
> '*cmusaslsecretCRAM-MD5'
> I0226 03:17:26.769629 1115 auxprop.cpp:107] Request to lookup properties for
> user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN:
> 'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true
> I0226 03:17:26.769637 1115 auxprop.cpp:129] Skipping auxiliary property
> '*userPassword' since SASL_AUXPROP_AUTHZID == true
> I0226 03:17:26.769642 1115 auxprop.cpp:129] Skipping auxiliary property
> '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true
> I0226 03:17:26.769654 1115 authenticator.cpp:317] Authentication success
> I0226 03:17:26.769728 1117 authenticatee.cpp:298] Authentication success
> I0226 03:17:26.769769 1112 authenticatee.cpp:212] Received SASL
> authentication mechanisms: CRAM-MD5
> I0226 03:17:26.769767 1118 master.cpp:5556] Successfully authenticated
> principal 'test-principal' at
> [email protected]:51934
> I0226 03:17:26.769803 1112 authenticatee.cpp:238] Attempting to authenticate
> with mechanism 'CRAM-MD5'
> I0226 03:17:26.769798 1114 authenticator.cpp:431] Authentication session
> cleanup for crammd5_authenticatee(839)@172.30.2.151:51934
> I0226 03:17:26.769881 1112 authenticator.cpp:203] Received SASL
> authentication start
> I0226 03:17:26.769932 1112 authenticator.cpp:325] Authentication requires
> more steps
> I0226 03:17:26.769981 1117 sched.cpp:471] Successfully authenticated with
> master [email protected]:51934
> I0226 03:17:26.770004 1117 sched.cpp:776] Sending SUBSCRIBE call to
> [email protected]:51934
> I0226 03:17:26.770064 1118 authenticatee.cpp:258] Received SASL
> authentication step
> I0226 03:17:26.770102 1117 sched.cpp:809] Will retry registration in
> 1.937819802secs if necessary
> I0226 03:17:26.770165 1115 authenticator.cpp:231] Received SASL
> authentication step
> I0226 03:17:26.770193 1115 auxprop.cpp:107] Request to lookup properties for
> user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN:
> 'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false
> I0226 03:17:26.770207 1115 auxprop.cpp:179] Looking up auxiliary property
> '*userPassword'
> I0226 03:17:26.770213 1116 master.cpp:2280] Received SUBSCRIBE call for
> framework 'default' at
> [email protected]:51934
> I0226 03:17:26.770241 1115 auxprop.cpp:179] Looking up auxiliary property
> '*cmusaslsecretCRAM-MD5'
> I0226 03:17:26.770274 1115 auxprop.cpp:107] Request to lookup properties for
> user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN:
> 'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true
> I0226 03:17:26.770277 1116 master.cpp:1751] Authorizing framework principal
> 'test-principal' to receive offers for role 'role1'
> I0226 03:17:26.770298 1115 auxprop.cpp:129] Skipping auxiliary property
> '*userPassword' since SASL_AUXPROP_AUTHZID == true
> I0226 03:17:26.770331 1115 auxprop.cpp:129] Skipping auxiliary property
> '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true
> I0226 03:17:26.770349 1115 authenticator.cpp:317] Authentication success
> I0226 03:17:26.770428 1118 authenticatee.cpp:298] Authentication success
> I0226 03:17:26.770442 1116 master.cpp:5556] Successfully authenticated
> principal 'test-principal' at slave(399)@172.30.2.151:51934
> I0226 03:17:26.770547 1116 authenticator.cpp:431] Authentication session
> cleanup for crammd5_authenticatee(840)@172.30.2.151:51934
> I0226 03:17:26.770846 1116 master.cpp:2351] Subscribing framework default
> with checkpointing enabled and capabilities [ ]
> I0226 03:17:26.770866 1118 slave.cpp:927] Successfully authenticated with
> master [email protected]:51934
> I0226 03:17:26.770966 1118 slave.cpp:1321] Will retry registration in
> 1.453415ms if necessary
> I0226 03:17:26.771225 1115 hierarchical.cpp:265] Added framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.771275 1118 sched.cpp:703] Framework registered with
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.771299 1115 hierarchical.cpp:1434] No resources available to
> allocate!
> I0226 03:17:26.771328 1115 hierarchical.cpp:1529] No inverse offers to send
> out!
> I0226 03:17:26.771344 1118 sched.cpp:717] Scheduler::registered took 50146ns
> I0226 03:17:26.771356 1116 master.cpp:4240] Registering slave at
> slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) with id
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
> I0226 03:17:26.771348 1115 hierarchical.cpp:1127] Performed allocation for 0
> slaves in 101438ns
> I0226 03:17:26.771860 1114 registrar.cpp:439] Applied 1 operations in
> 59672ns; attempting to update the 'registry'
> I0226 03:17:26.772645 1117 log.cpp:683] Attempting to append 423 bytes to
> the log
> I0226 03:17:26.772758 1112 coordinator.cpp:348] Coordinator attempting to
> write APPEND action at position 3
> I0226 03:17:26.773435 1117 replica.cpp:537] Replica received write request
> for position 3 from (13824)@172.30.2.151:51934
> I0226 03:17:26.773586 1111 slave.cpp:1321] Will retry registration in
> 2.74261ms if necessary
> I0226 03:17:26.773682 1115 master.cpp:4228] Ignoring register slave message
> from slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) as
> admission is already in progress
> I0226 03:17:26.773937 1117 leveldb.cpp:341] Persisting action (442 bytes) to
> leveldb took 469969ns
> I0226 03:17:26.773957 1117 replica.cpp:712] Persisted action at 3
> I0226 03:17:26.774605 1114 replica.cpp:691] Replica received learned notice
> for position 3 from @0.0.0.0:0
> I0226 03:17:26.775961 1114 leveldb.cpp:341] Persisting action (444 bytes) to
> leveldb took 1.329435ms
> I0226 03:17:26.775986 1114 replica.cpp:712] Persisted action at 3
> I0226 03:17:26.776008 1114 replica.cpp:697] Replica learned APPEND action at
> position 3
> I0226 03:17:26.777228 1115 slave.cpp:1321] Will retry registration in
> 41.5608ms if necessary
> I0226 03:17:26.777300 1112 registrar.cpp:484] Successfully updated the
> 'registry' in 5.378048ms
> I0226 03:17:26.777361 1114 master.cpp:4228] Ignoring register slave message
> from slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) as
> admission is already in progress
> I0226 03:17:26.777505 1113 log.cpp:702] Attempting to truncate the log to 3
> I0226 03:17:26.777616 1111 coordinator.cpp:348] Coordinator attempting to
> write TRUNCATE action at position 4
> I0226 03:17:26.778062 1114 slave.cpp:3482] Received ping from
> slave-observer(369)@172.30.2.151:51934
> I0226 03:17:26.778139 1118 master.cpp:4308] Registered slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io) with cpu(*):2; mem(*):2048; disk(role1):2048;
> cpus(*):8; ports(*):[31000-32000]
> I0226 03:17:26.778213 1113 replica.cpp:537] Replica received write request
> for position 4 from (13825)@172.30.2.151:51934
> I0226 03:17:26.778291 1114 slave.cpp:971] Registered with master
> [email protected]:51934; given slave ID
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
> I0226 03:17:26.778316 1114 fetcher.cpp:81] Clearing fetcher cache
> I0226 03:17:26.778367 1116 hierarchical.cpp:473] Added slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io) with
> cpu(*):2; mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000]
> (allocated: )
> I0226 03:17:26.778447 1117 status_update_manager.cpp:181] Resuming sending
> status updates
> I0226 03:17:26.778617 1113 leveldb.cpp:341] Persisting action (16 bytes) to
> leveldb took 375414ns
> I0226 03:17:26.778635 1113 replica.cpp:712] Persisted action at 4
> I0226 03:17:26.778650 1114 slave.cpp:994] Checkpointing SlaveInfo to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/slave.info'
> I0226 03:17:26.778900 1114 slave.cpp:1030] Forwarding total oversubscribed
> resources
> I0226 03:17:26.779109 1114 master.cpp:4649] Received update of slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io) with total oversubscribed resources
> I0226 03:17:26.779139 1112 replica.cpp:691] Replica received learned notice
> for position 4 from @0.0.0.0:0
> I0226 03:17:26.779331 1116 hierarchical.cpp:1529] No inverse offers to send
> out!
> I0226 03:17:26.779369 1116 hierarchical.cpp:1147] Performed allocation for
> slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 in 969593ns
> I0226 03:17:26.779645 1113 master.cpp:5355] Sending 1 offers to framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at
> [email protected]:51934
> I0226 03:17:26.779700 1116 hierarchical.cpp:531] Slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io)
> updated with oversubscribed resources (total: cpu(*):2; mem(*):2048;
> disk(role1):2048; cpus(*):8; ports(*):[31000-32000], allocated:
> disk(role1):2048; cpu(*):2; mem(*):2048; cpus(*):8; ports(*):[31000-32000])
> I0226 03:17:26.779819 1116 hierarchical.cpp:1434] No resources available to
> allocate!
> I0226 03:17:26.779847 1116 hierarchical.cpp:1529] No inverse offers to send
> out!
> I0226 03:17:26.779865 1116 hierarchical.cpp:1147] Performed allocation for
> slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 in 133437ns
> I0226 03:17:26.780025 1118 sched.cpp:873] Scheduler::resourceOffers took
> 102165ns
> I0226 03:17:26.780372 1097 resources.cpp:576] Parsing resources as JSON
> failed: cpus:1;mem:64;
> Trying semicolon-delimited string format instead
> I0226 03:17:26.780882 1112 leveldb.cpp:341] Persisting action (18 bytes) to
> leveldb took 1.715066ms
> I0226 03:17:26.780938 1112 leveldb.cpp:399] Deleting ~2 keys from leveldb
> took 32370ns
> I0226 03:17:26.780953 1112 replica.cpp:712] Persisted action at 4
> I0226 03:17:26.780971 1112 replica.cpp:697] Replica learned TRUNCATE action
> at position 4
> I0226 03:17:26.781693 1117 master.cpp:3138] Processing ACCEPT call for
> offers: [ 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-O0 ] on slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io) for framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at
> [email protected]:51934
> I0226 03:17:26.781731 1117 master.cpp:2926] Authorizing principal
> 'test-principal' to create volumes
> I0226 03:17:26.781801 1117 master.cpp:2825] Authorizing framework principal
> 'test-principal' to launch task 1 as user 'root'
> I0226 03:17:26.782827 1114 master.cpp:3467] Applying CREATE operation for
> volumes disk(role1)[id1:path1]:64 from framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at
> [email protected]:51934 to slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:26.783136 1114 master.cpp:6589] Sending checkpointed resources
> disk(role1)[id1:path1]:64 to slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at
> slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:26.783641 1111 slave.cpp:2341] Updated checkpointed resources
> from to disk(role1)[id1:path1]:64
> I0226 03:17:26.783911 1114 master.hpp:176] Adding task 1 with resources
> cpus(*):1; mem(*):64; disk(role1)[id1:path1]:64 on slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:26.784056 1114 master.cpp:3623] Launching task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at
> [email protected]:51934 with
> resources cpus(*):1; mem(*):64; disk(role1)[id1:path1]:64 on slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:26.784397 1115 slave.cpp:1361] Got assigned task 1 for framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.784557 1115 slave.cpp:5287] Checkpointing FrameworkInfo to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/framework.info'
> I0226 03:17:26.784739 1116 hierarchical.cpp:653] Updated allocation of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 on slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 from disk(role1):2048; cpu(*):2;
> mem(*):2048; cpus(*):8; ports(*):[31000-32000] to disk(role1):1984; cpu(*):2;
> mem(*):2048; cpus(*):8; ports(*):[31000-32000]; disk(role1)[id1:path1]:64
> I0226 03:17:26.784848 1115 slave.cpp:5298] Checkpointing framework pid
> '[email protected]:51934' to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/framework.pid'
> I0226 03:17:26.785078 1115 resources.cpp:576] Parsing resources as JSON
> failed: cpus:0.1;mem:32
> Trying semicolon-delimited string format instead
> I0226 03:17:26.785322 1116 hierarchical.cpp:892] Recovered disk(role1):1984;
> cpu(*):2; mem(*):1984; cpus(*):7; ports(*):[31000-32000] (total: cpu(*):2;
> mem(*):2048; disk(role1):1984; cpus(*):8; ports(*):[31000-32000];
> disk(role1)[id1:path1]:64, allocated: disk(role1)[id1:path1]:64; cpus(*):1;
> mem(*):64) on slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 from framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.785658 1115 slave.cpp:1480] Launching task 1 for framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.785719 1115 resources.cpp:576] Parsing resources as JSON
> failed: cpus:0.1;mem:32
> Trying semicolon-delimited string format instead
> I0226 03:17:26.786197 1115 paths.cpp:474] Trying to chown
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> to user 'root'
> I0226 03:17:26.791122 1115 slave.cpp:5739] Checkpointing ExecutorInfo to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/executor.info'
> I0226 03:17:26.791543 1115 slave.cpp:5367] Launching executor 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 with resources cpus(*):0.1;
> mem(*):32 in work directory
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> I0226 03:17:26.792325 1115 slave.cpp:5762] Checkpointing TaskInfo to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/tasks/1/task.info'
> I0226 03:17:26.794337 1115 slave.cpp:1698] Queuing task '1' for executor '1'
> of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:26.794478 1115 slave.cpp:749] Successfully attached file
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> I0226 03:17:26.797106 1116 docker.cpp:1023] Starting container
> 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' for task '1' (and executor '1') of
> framework '5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000'
> I0226 03:17:26.797462 1116 docker.cpp:1053] Running docker -H
> unix:///var/run/docker.sock inspect alpine:latest
> I0226 03:17:26.910549 1111 docker.cpp:394] Docker pull alpine completed
> I0226 03:17:26.910800 1111 docker.cpp:483] Changing the ownership of the
> persistent volume at
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/volumes/roles/role1/id1'
> with uid 0 and gid 0
> I0226 03:17:26.915712 1111 docker.cpp:504] Mounting
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/volumes/roles/role1/id1'
> to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1'
> for persistent volume disk(role1)[id1:path1]:64 of container
> bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c
> I0226 03:17:26.919000 1117 docker.cpp:576] Checkpointing pid 9568 to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/pids/forked.pid'
> I0226 03:17:26.974776 1114 slave.cpp:2643] Got registration for executor '1'
> of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from
> executor(1)@172.30.2.151:46052
> I0226 03:17:26.975217 1114 slave.cpp:2729] Checkpointing executor pid
> 'executor(1)@172.30.2.151:46052' to
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/pids/libprocess.pid'
> I0226 03:17:26.976177 1113 docker.cpp:1303] Ignoring updating container
> 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' with resources passed to update is
> identical to existing resources
> I0226 03:17:26.976492 1115 slave.cpp:1863] Sending queued task '1' to
> executor '1' of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 at
> executor(1)@172.30.2.151:46052
> I0226 03:17:27.691769 1111 slave.cpp:3002] Handling status update
> TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from
> executor(1)@172.30.2.151:46052
> I0226 03:17:27.692291 1116 status_update_manager.cpp:320] Received status
> update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1
> of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.692327 1116 status_update_manager.cpp:497] Creating
> StatusUpdate stream for task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.692773 1116 status_update_manager.cpp:824] Checkpointing
> UPDATE for status update TASK_RUNNING (UUID:
> 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.700090 1116 status_update_manager.cpp:374] Forwarding update
> TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to the slave
> I0226 03:17:27.700389 1113 slave.cpp:3400] Forwarding the update
> TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to
> [email protected]:51934
> I0226 03:17:27.700606 1113 slave.cpp:3294] Status update manager
> successfully handled status update TASK_RUNNING (UUID:
> 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.700644 1113 slave.cpp:3310] Sending acknowledgement for
> status update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for
> task 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to
> executor(1)@172.30.2.151:46052
> I0226 03:17:27.700742 1117 master.cpp:4794] Status update TASK_RUNNING
> (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:27.700775 1117 master.cpp:4842] Forwarding status update
> TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.700923 1117 master.cpp:6450] Updating the state of task 1 of
> framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (latest state:
> TASK_RUNNING, status update state: TASK_RUNNING)
> I0226 03:17:27.701145 1118 sched.cpp:981] Scheduler::statusUpdate took
> 107222ns
> I0226 03:17:27.701550 1112 master.cpp:3952] Processing ACKNOWLEDGE call
> 9f75a4e5-9ff4-4ca9-8623-8b2574796229 for task 1 of framework
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at
> [email protected]:51934 on slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
> I0226 03:17:27.701828 1114 status_update_manager.cpp:392] Received status
> update acknowledgement (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task
> 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.701962 1114 status_update_manager.cpp:824] Checkpointing ACK
> for status update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229)
> for task 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
> I0226 03:17:27.701987 1112 slave.cpp:668] Slave terminating
> I0226 03:17:27.702256 1117 master.cpp:1174] Slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io) disconnected
> I0226 03:17:27.702275 1117 master.cpp:2635] Disconnecting slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:27.702335 1117 master.cpp:2654] Deactivating slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934
> (ip-172-30-2-151.mesosphere.io)
> I0226 03:17:27.702492 1111 hierarchical.cpp:560] Slave
> 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 deactivated
> I0226 03:17:27.707713 1115 slave.cpp:193] Slave started on
> 400)@172.30.2.151:51934
> I0226 03:17:27.707739 1115 slave.cpp:194] Flags at startup:
> --appc_simple_discovery_uri_prefix="http://"
> --appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5"
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false"
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false"
> --cgroups_root="mesos" --container_disk_watch_interval="15secs"
> --containerizers="mesos"
> --credential="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential"
> --default_role="*" --disk_watch_interval="1mins" --docker="docker"
> --docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io"
> --docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock"
> --docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker"
> --enforce_container_disk_quota="false"
> --executor_registration_timeout="1mins"
> --executor_shutdown_grace_period="5secs"
> --fetcher_cache_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/fetch"
> --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks"
> --gc_disk_headroom="0.1" --hadoop_home="" --help="false"
> --hostname_lookup="true" --image_provisioner_backend="copy"
> --initialize_driver_logging="true" --isolation="posix/cpu,posix/mem"
> --launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src"
> --logbufsecs="0" --logging_level="INFO"
> --oversubscribed_resources_interval="15secs" --perf_duration="10secs"
> --perf_interval="1mins" --qos_correction_interval_min="0ns" --quiet="false"
> --recover="reconnect" --recovery_timeout="15mins"
> --registration_backoff_factor="10ms"
> --resources="cpu:2;mem:2048;disk(role1):2048"
> --revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox"
> --strict="true" --switch_user="true" --systemd_enable_support="true"
> --systemd_runtime_directory="/run/systemd/system" --version="false"
> --work_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP"
> I0226 03:17:27.708133 1115 credentials.hpp:83] Loading credential for
> authentication from
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential'
> I0226 03:17:27.708282 1115 slave.cpp:324] Slave using credential for:
> test-principal
> I0226 03:17:27.708407 1115 resources.cpp:576] Parsing resources as JSON
> failed: cpu:2;mem:2048;disk(role1):2048
> Trying semicolon-delimited string format instead
> I0226 03:17:27.708874 1115 slave.cpp:464] Slave resources: cpu(*):2;
> mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000]
> I0226 03:17:27.708931 1115 slave.cpp:472] Slave attributes: [ ]
> I0226 03:17:27.708941 1115 slave.cpp:477] Slave hostname:
> ip-172-30-2-151.mesosphere.io
> I0226 03:17:27.710033 1113 state.cpp:58] Recovering state from
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta'
> I0226 03:17:27.711252 1114 fetcher.cpp:81] Clearing fetcher cache
> I0226 03:17:27.711447 1116 status_update_manager.cpp:200] Recovering status
> update manager
> I0226 03:17:27.711727 1111 docker.cpp:726] Recovering Docker containers
> I0226 03:17:27.711839 1111 docker.cpp:885] Running docker -H
> unix:///var/run/docker.sock ps -a
> I0226 03:17:27.728170 1117 hierarchical.cpp:1434] No resources available to
> allocate!
> I0226 03:17:27.728235 1117 hierarchical.cpp:1529] No inverse offers to send
> out!
> I0226 03:17:27.728268 1117 hierarchical.cpp:1127] Performed allocation for 1
> slaves in 296715ns
> I0226 03:17:27.817551 1113 docker.cpp:766] Running docker -H
> unix:///var/run/docker.sock inspect
> mesos-5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0.bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c
> I0226 03:17:27.923014 1112 docker.cpp:932] Checking if Docker container
> named
> '/mesos-5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0.bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> was started by Mesos
> I0226 03:17:27.923071 1112 docker.cpp:942] Checking if Mesos container with
> ID 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' has been orphaned
> I0226 03:17:27.923122 1112 docker.cpp:678] Running docker -H
> unix:///var/run/docker.sock stop -t 0
> 0a10ad8641f8e85227324a979817933322dc901706cb4430eab0bcaf979835d1
> I0226 03:17:28.023885 1116 docker.cpp:727] Running docker -H
> unix:///var/run/docker.sock rm -v
> 0a10ad8641f8e85227324a979817933322dc901706cb4430eab0bcaf979835d1
> I0226 03:17:28.127876 1114 docker.cpp:912] Unmounting volume for container
> 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> ../../3rdparty/libprocess/include/process/gmock.hpp:214: ERROR: this mock
> object (used in test
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should
> be deleted but never is. Its address is @0x5781dd8.
> I0226 03:17:28.127957 1114 docker.cpp:912] Unmounting volume for container
> 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
> ../../src/tests/mesos.cpp:673: ERROR: this mock object (used in test
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should
> be deleted but never is. Its address is @0x5a03260.
> ../../src/tests/mesos.hpp:1357: ERROR: this mock object (used in test
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should
> be deleted but never is. Its address is @0x5b477c0.
> Failed to perform recovery: Unable to unmount volumes for Docker container
> 'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c': Failed to unmount volume
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1':
> Failed to unmount
> '/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1':
> Invalid argument
> ../../src/tests/containerizer/docker_containerizer_tests.cpp:1650: ERROR:
> this mock object (used in test
> DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should
> be deleted but never is. Its address is @0x7ffe75a8d310.
> To remedy this do as follows:
> ERROR: 4 leaked mock objects found at program exit.
> Step 1: rm -f
> /tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/latest
> This ensures slave doesn't recover old live executors.
> Step 2: Restart the slave.
> Process exited with code 1
> {code}
> There appear to be two problems:
> 1) The docker containerizer should not exit on failure to clean up orphans.
> The MesosContainerizer does not do this (see [MESOS-2367]).
> 2) Unmounting the orphan persistent volume fails for some reason.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)