Joseph Wu created MESOS-4832:
--------------------------------

             Summary: 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes exits when 
the /tmp directory is bind-mounted
                 Key: MESOS-4832
                 URL: https://issues.apache.org/jira/browse/MESOS-4832
             Project: Mesos
          Issue Type: Bug
          Components: containerization, docker
    Affects Versions: 0.27.0
         Environment: Seen on CentOS 7 & Debian 8.
            Reporter: Joseph Wu
            Assignee: Jie Yu


If the {{/tmp}} directory (where Mesos tests create temporary directories) is a 
bind mount, the test suite will exit here:
{code}
[ RUN      ] 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes
I0226 03:17:26.722806  1097 leveldb.cpp:174] Opened db in 12.587676ms
I0226 03:17:26.723496  1097 leveldb.cpp:181] Compacted db in 636999ns
I0226 03:17:26.723536  1097 leveldb.cpp:196] Created db iterator in 18271ns
I0226 03:17:26.723547  1097 leveldb.cpp:202] Seeked to beginning of db in 1555ns
I0226 03:17:26.723554  1097 leveldb.cpp:271] Iterated through 0 keys in the db 
in 363ns
I0226 03:17:26.723593  1097 replica.cpp:779] Replica recovered with log 
positions 0 -> 0 with 1 holes and 0 unlearned
I0226 03:17:26.724128  1117 recover.cpp:447] Starting replica recovery
I0226 03:17:26.724367  1117 recover.cpp:473] Replica is in EMPTY status
I0226 03:17:26.725237  1117 replica.cpp:673] Replica in EMPTY status received a 
broadcasted recover request from (13810)@172.30.2.151:51934
I0226 03:17:26.725744  1114 recover.cpp:193] Received a recover response from a 
replica in EMPTY status
I0226 03:17:26.726356  1111 master.cpp:376] Master 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a (ip-172-30-2-151.mesosphere.io) started on 
172.30.2.151:51934
I0226 03:17:26.726369  1118 recover.cpp:564] Updating replica status to STARTING
I0226 03:17:26.726378  1111 master.cpp:378] Flags at startup: --acls="" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate="true" --authenticate_http="true" --authenticate_slaves="true" 
--authenticators="crammd5" --authorizers="local" 
--credentials="/tmp/djHTVQ/credentials" --framework_sorter="drf" --help="false" 
--hostname_lookup="true" --http_authenticators="basic" 
--initialize_driver_logging="true" --log_auto_initialize="true" 
--logbufsecs="0" --logging_level="INFO" --max_completed_frameworks="50" 
--max_completed_tasks_per_framework="1000" --max_slave_ping_timeouts="5" 
--quiet="false" --recovery_slave_removal_limit="100%" 
--registry="replicated_log" --registry_fetch_timeout="1mins" 
--registry_store_timeout="100secs" --registry_strict="true" 
--root_submissions="true" --slave_ping_timeout="15secs" 
--slave_reregister_timeout="10mins" --user_sorter="drf" --version="false" 
--webui_dir="/usr/local/share/mesos/webui" --work_dir="/tmp/djHTVQ/master" 
--zk_session_timeout="10secs"
I0226 03:17:26.726605  1111 master.cpp:423] Master only allowing authenticated 
frameworks to register
I0226 03:17:26.726616  1111 master.cpp:428] Master only allowing authenticated 
slaves to register
I0226 03:17:26.726632  1111 credentials.hpp:35] Loading credentials for 
authentication from '/tmp/djHTVQ/credentials'
I0226 03:17:26.726860  1111 master.cpp:468] Using default 'crammd5' 
authenticator
I0226 03:17:26.726977  1111 master.cpp:537] Using default 'basic' HTTP 
authenticator
I0226 03:17:26.727092  1111 master.cpp:571] Authorization enabled
I0226 03:17:26.727243  1118 hierarchical.cpp:144] Initialized hierarchical 
allocator process
I0226 03:17:26.727285  1116 whitelist_watcher.cpp:77] No whitelist given
I0226 03:17:26.728852  1114 master.cpp:1712] The newly elected leader is 
master@172.30.2.151:51934 with id 5cc57c0e-f1ad-4107-893f-420ed1a1db1a
I0226 03:17:26.728876  1114 master.cpp:1725] Elected as the leading master!
I0226 03:17:26.728891  1114 master.cpp:1470] Recovering from registrar
I0226 03:17:26.728977  1117 registrar.cpp:307] Recovering registrar
I0226 03:17:26.731503  1112 leveldb.cpp:304] Persisting metadata (8 bytes) to 
leveldb took 4.977811ms
I0226 03:17:26.731539  1112 replica.cpp:320] Persisted replica status to 
STARTING
I0226 03:17:26.731711  1111 recover.cpp:473] Replica is in STARTING status
I0226 03:17:26.732501  1114 replica.cpp:673] Replica in STARTING status 
received a broadcasted recover request from (13812)@172.30.2.151:51934
I0226 03:17:26.732862  1111 recover.cpp:193] Received a recover response from a 
replica in STARTING status
I0226 03:17:26.733264  1117 recover.cpp:564] Updating replica status to VOTING
I0226 03:17:26.733836  1118 leveldb.cpp:304] Persisting metadata (8 bytes) to 
leveldb took 388246ns
I0226 03:17:26.733855  1118 replica.cpp:320] Persisted replica status to VOTING
I0226 03:17:26.733979  1113 recover.cpp:578] Successfully joined the Paxos group
I0226 03:17:26.734149  1113 recover.cpp:462] Recover process terminated
I0226 03:17:26.734478  1111 log.cpp:659] Attempting to start the writer
I0226 03:17:26.735523  1114 replica.cpp:493] Replica received implicit promise 
request from (13813)@172.30.2.151:51934 with proposal 1
I0226 03:17:26.736130  1114 leveldb.cpp:304] Persisting metadata (8 bytes) to 
leveldb took 576451ns
I0226 03:17:26.736150  1114 replica.cpp:342] Persisted promised to 1
I0226 03:17:26.736709  1115 coordinator.cpp:238] Coordinator attempting to fill 
missing positions
I0226 03:17:26.737771  1114 replica.cpp:388] Replica received explicit promise 
request from (13814)@172.30.2.151:51934 for position 0 with proposal 2
I0226 03:17:26.738386  1114 leveldb.cpp:341] Persisting action (8 bytes) to 
leveldb took 583184ns
I0226 03:17:26.738404  1114 replica.cpp:712] Persisted action at 0
I0226 03:17:26.739312  1118 replica.cpp:537] Replica received write request for 
position 0 from (13815)@172.30.2.151:51934
I0226 03:17:26.739367  1118 leveldb.cpp:436] Reading position from leveldb took 
26157ns
I0226 03:17:26.740638  1118 leveldb.cpp:341] Persisting action (14 bytes) to 
leveldb took 1.238477ms
I0226 03:17:26.740669  1118 replica.cpp:712] Persisted action at 0
I0226 03:17:26.741158  1118 replica.cpp:691] Replica received learned notice 
for position 0 from @0.0.0.0:0
I0226 03:17:26.742878  1118 leveldb.cpp:341] Persisting action (16 bytes) to 
leveldb took 1.697254ms
I0226 03:17:26.742902  1118 replica.cpp:712] Persisted action at 0
I0226 03:17:26.742916  1118 replica.cpp:697] Replica learned NOP action at 
position 0
I0226 03:17:26.743393  1117 log.cpp:675] Writer started with ending position 0
I0226 03:17:26.744370  1112 leveldb.cpp:436] Reading position from leveldb took 
34329ns
I0226 03:17:26.745240  1117 registrar.cpp:340] Successfully fetched the 
registry (0B) in 16.21888ms
I0226 03:17:26.745350  1117 registrar.cpp:439] Applied 1 operations in 30460ns; 
attempting to update the 'registry'
I0226 03:17:26.746016  1111 log.cpp:683] Attempting to append 210 bytes to the 
log
I0226 03:17:26.746119  1116 coordinator.cpp:348] Coordinator attempting to 
write APPEND action at position 1
I0226 03:17:26.746798  1114 replica.cpp:537] Replica received write request for 
position 1 from (13816)@172.30.2.151:51934
I0226 03:17:26.747251  1114 leveldb.cpp:341] Persisting action (229 bytes) to 
leveldb took 411333ns
I0226 03:17:26.747269  1114 replica.cpp:712] Persisted action at 1
I0226 03:17:26.747808  1113 replica.cpp:691] Replica received learned notice 
for position 1 from @0.0.0.0:0
I0226 03:17:26.749511  1113 leveldb.cpp:341] Persisting action (231 bytes) to 
leveldb took 1.673488ms
I0226 03:17:26.749534  1113 replica.cpp:712] Persisted action at 1
I0226 03:17:26.749550  1113 replica.cpp:697] Replica learned APPEND action at 
position 1
I0226 03:17:26.750422  1111 registrar.cpp:484] Successfully updated the 
'registry' in 5.021952ms
I0226 03:17:26.750560  1111 registrar.cpp:370] Successfully recovered registrar
I0226 03:17:26.750635  1112 log.cpp:702] Attempting to truncate the log to 1
I0226 03:17:26.750751  1113 coordinator.cpp:348] Coordinator attempting to 
write TRUNCATE action at position 2
I0226 03:17:26.751096  1116 master.cpp:1522] Recovered 0 slaves from the 
Registry (171B) ; allowing 10mins for slaves to re-register
I0226 03:17:26.751126  1111 hierarchical.cpp:171] Skipping recovery of 
hierarchical allocator: nothing to recover
I0226 03:17:26.751561  1118 replica.cpp:537] Replica received write request for 
position 2 from (13817)@172.30.2.151:51934
I0226 03:17:26.751999  1118 leveldb.cpp:341] Persisting action (16 bytes) to 
leveldb took 406823ns
I0226 03:17:26.752018  1118 replica.cpp:712] Persisted action at 2
I0226 03:17:26.752521  1113 replica.cpp:691] Replica received learned notice 
for position 2 from @0.0.0.0:0
I0226 03:17:26.754161  1113 leveldb.cpp:341] Persisting action (18 bytes) to 
leveldb took 1.614888ms
I0226 03:17:26.754210  1113 leveldb.cpp:399] Deleting ~1 keys from leveldb took 
26384ns
I0226 03:17:26.754225  1113 replica.cpp:712] Persisted action at 2
I0226 03:17:26.754240  1113 replica.cpp:697] Replica learned TRUNCATE action at 
position 2
I0226 03:17:26.765103  1115 slave.cpp:193] Slave started on 
399)@172.30.2.151:51934
I0226 03:17:26.765130  1115 slave.cpp:194] Flags at startup: 
--appc_simple_discovery_uri_prefix="http://"; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" 
--credential="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential"
 --default_role="*" --disk_watch_interval="1mins" --docker="docker" 
--docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io"; 
--docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock" 
--docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" 
--fetcher_cache_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/fetch"
 --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname_lookup="true" --image_provisioner_backend="copy" 
--initialize_driver_logging="true" --isolation="posix/cpu,posix/mem" 
--launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src" --logbufsecs="0" 
--logging_level="INFO" --oversubscribed_resources_interval="15secs" 
--perf_duration="10secs" --perf_interval="1mins" 
--qos_correction_interval_min="0ns" --quiet="false" --recover="reconnect" 
--recovery_timeout="15mins" --registration_backoff_factor="10ms" 
--resources="cpu:2;mem:2048;disk(role1):2048" 
--revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox" 
--strict="true" --switch_user="true" --systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP"
I0226 03:17:26.765403  1115 credentials.hpp:83] Loading credential for 
authentication from 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential'
I0226 03:17:26.765573  1115 slave.cpp:324] Slave using credential for: 
test-principal
I0226 03:17:26.765733  1115 resources.cpp:576] Parsing resources as JSON 
failed: cpu:2;mem:2048;disk(role1):2048
Trying semicolon-delimited string format instead
I0226 03:17:26.766185  1115 slave.cpp:464] Slave resources: cpu(*):2; 
mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000]
I0226 03:17:26.766242  1115 slave.cpp:472] Slave attributes: [  ]
I0226 03:17:26.766250  1115 slave.cpp:477] Slave hostname: 
ip-172-30-2-151.mesosphere.io
I0226 03:17:26.767325  1097 sched.cpp:222] Version: 0.28.0
I0226 03:17:26.767390  1111 state.cpp:58] Recovering state from 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta'
I0226 03:17:26.767603  1115 status_update_manager.cpp:200] Recovering status 
update manager
I0226 03:17:26.767865  1113 docker.cpp:726] Recovering Docker containers
I0226 03:17:26.767971  1111 sched.cpp:326] New master detected at 
master@172.30.2.151:51934
I0226 03:17:26.768045  1111 sched.cpp:382] Authenticating with master 
master@172.30.2.151:51934
I0226 03:17:26.768059  1111 sched.cpp:389] Using default CRAM-MD5 authenticatee
I0226 03:17:26.768070  1118 slave.cpp:4565] Finished recovery
I0226 03:17:26.768273  1112 authenticatee.cpp:121] Creating new client SASL 
connection
I0226 03:17:26.768435  1118 slave.cpp:4737] Querying resource estimator for 
oversubscribable resources
I0226 03:17:26.768565  1111 master.cpp:5526] Authenticating 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934
I0226 03:17:26.768661  1118 slave.cpp:796] New master detected at 
master@172.30.2.151:51934
I0226 03:17:26.768659  1115 authenticator.cpp:413] Starting authentication 
session for crammd5_authenticatee(839)@172.30.2.151:51934
I0226 03:17:26.768679  1113 status_update_manager.cpp:174] Pausing sending 
status updates
I0226 03:17:26.768728  1118 slave.cpp:859] Authenticating with master 
master@172.30.2.151:51934
I0226 03:17:26.768743  1118 slave.cpp:864] Using default CRAM-MD5 authenticatee
I0226 03:17:26.768865  1118 slave.cpp:832] Detecting new master
I0226 03:17:26.768868  1112 authenticator.cpp:98] Creating new server SASL 
connection
I0226 03:17:26.768908  1114 authenticatee.cpp:121] Creating new client SASL 
connection
I0226 03:17:26.769003  1118 slave.cpp:4751] Received oversubscribable resources 
 from the resource estimator
I0226 03:17:26.769103  1115 authenticatee.cpp:212] Received SASL authentication 
mechanisms: CRAM-MD5
I0226 03:17:26.769131  1115 authenticatee.cpp:238] Attempting to authenticate 
with mechanism 'CRAM-MD5'
I0226 03:17:26.769209  1116 master.cpp:5526] Authenticating 
slave(399)@172.30.2.151:51934
I0226 03:17:26.769253  1114 authenticator.cpp:203] Received SASL authentication 
start
I0226 03:17:26.769295  1115 authenticator.cpp:413] Starting authentication 
session for crammd5_authenticatee(840)@172.30.2.151:51934
I0226 03:17:26.769307  1114 authenticator.cpp:325] Authentication requires more 
steps
I0226 03:17:26.769403  1117 authenticatee.cpp:258] Received SASL authentication 
step
I0226 03:17:26.769495  1114 authenticator.cpp:98] Creating new server SASL 
connection
I0226 03:17:26.769531  1115 authenticator.cpp:231] Received SASL authentication 
step
I0226 03:17:26.769554  1115 auxprop.cpp:107] Request to lookup properties for 
user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN: 
'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false 
I0226 03:17:26.769562  1115 auxprop.cpp:179] Looking up auxiliary property 
'*userPassword'
I0226 03:17:26.769608  1115 auxprop.cpp:179] Looking up auxiliary property 
'*cmusaslsecretCRAM-MD5'
I0226 03:17:26.769629  1115 auxprop.cpp:107] Request to lookup properties for 
user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN: 
'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true 
I0226 03:17:26.769637  1115 auxprop.cpp:129] Skipping auxiliary property 
'*userPassword' since SASL_AUXPROP_AUTHZID == true
I0226 03:17:26.769642  1115 auxprop.cpp:129] Skipping auxiliary property 
'*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true
I0226 03:17:26.769654  1115 authenticator.cpp:317] Authentication success
I0226 03:17:26.769728  1117 authenticatee.cpp:298] Authentication success
I0226 03:17:26.769769  1112 authenticatee.cpp:212] Received SASL authentication 
mechanisms: CRAM-MD5
I0226 03:17:26.769767  1118 master.cpp:5556] Successfully authenticated 
principal 'test-principal' at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934
I0226 03:17:26.769803  1112 authenticatee.cpp:238] Attempting to authenticate 
with mechanism 'CRAM-MD5'
I0226 03:17:26.769798  1114 authenticator.cpp:431] Authentication session 
cleanup for crammd5_authenticatee(839)@172.30.2.151:51934
I0226 03:17:26.769881  1112 authenticator.cpp:203] Received SASL authentication 
start
I0226 03:17:26.769932  1112 authenticator.cpp:325] Authentication requires more 
steps
I0226 03:17:26.769981  1117 sched.cpp:471] Successfully authenticated with 
master master@172.30.2.151:51934
I0226 03:17:26.770004  1117 sched.cpp:776] Sending SUBSCRIBE call to 
master@172.30.2.151:51934
I0226 03:17:26.770064  1118 authenticatee.cpp:258] Received SASL authentication 
step
I0226 03:17:26.770102  1117 sched.cpp:809] Will retry registration in 
1.937819802secs if necessary
I0226 03:17:26.770165  1115 authenticator.cpp:231] Received SASL authentication 
step
I0226 03:17:26.770193  1115 auxprop.cpp:107] Request to lookup properties for 
user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN: 
'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false 
I0226 03:17:26.770207  1115 auxprop.cpp:179] Looking up auxiliary property 
'*userPassword'
I0226 03:17:26.770213  1116 master.cpp:2280] Received SUBSCRIBE call for 
framework 'default' at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934
I0226 03:17:26.770241  1115 auxprop.cpp:179] Looking up auxiliary property 
'*cmusaslsecretCRAM-MD5'
I0226 03:17:26.770274  1115 auxprop.cpp:107] Request to lookup properties for 
user: 'test-principal' realm: 'ip-172-30-2-151.mesosphere.io' server FQDN: 
'ip-172-30-2-151.mesosphere.io' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true 
I0226 03:17:26.770277  1116 master.cpp:1751] Authorizing framework principal 
'test-principal' to receive offers for role 'role1'
I0226 03:17:26.770298  1115 auxprop.cpp:129] Skipping auxiliary property 
'*userPassword' since SASL_AUXPROP_AUTHZID == true
I0226 03:17:26.770331  1115 auxprop.cpp:129] Skipping auxiliary property 
'*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true
I0226 03:17:26.770349  1115 authenticator.cpp:317] Authentication success
I0226 03:17:26.770428  1118 authenticatee.cpp:298] Authentication success
I0226 03:17:26.770442  1116 master.cpp:5556] Successfully authenticated 
principal 'test-principal' at slave(399)@172.30.2.151:51934
I0226 03:17:26.770547  1116 authenticator.cpp:431] Authentication session 
cleanup for crammd5_authenticatee(840)@172.30.2.151:51934
I0226 03:17:26.770846  1116 master.cpp:2351] Subscribing framework default with 
checkpointing enabled and capabilities [  ]
I0226 03:17:26.770866  1118 slave.cpp:927] Successfully authenticated with 
master master@172.30.2.151:51934
I0226 03:17:26.770966  1118 slave.cpp:1321] Will retry registration in 
1.453415ms if necessary
I0226 03:17:26.771225  1115 hierarchical.cpp:265] Added framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.771275  1118 sched.cpp:703] Framework registered with 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.771299  1115 hierarchical.cpp:1434] No resources available to 
allocate!
I0226 03:17:26.771328  1115 hierarchical.cpp:1529] No inverse offers to send 
out!
I0226 03:17:26.771344  1118 sched.cpp:717] Scheduler::registered took 50146ns
I0226 03:17:26.771356  1116 master.cpp:4240] Registering slave at 
slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) with id 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
I0226 03:17:26.771348  1115 hierarchical.cpp:1127] Performed allocation for 0 
slaves in 101438ns
I0226 03:17:26.771860  1114 registrar.cpp:439] Applied 1 operations in 59672ns; 
attempting to update the 'registry'
I0226 03:17:26.772645  1117 log.cpp:683] Attempting to append 423 bytes to the 
log
I0226 03:17:26.772758  1112 coordinator.cpp:348] Coordinator attempting to 
write APPEND action at position 3
I0226 03:17:26.773435  1117 replica.cpp:537] Replica received write request for 
position 3 from (13824)@172.30.2.151:51934
I0226 03:17:26.773586  1111 slave.cpp:1321] Will retry registration in 
2.74261ms if necessary
I0226 03:17:26.773682  1115 master.cpp:4228] Ignoring register slave message 
from slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) as admission 
is already in progress
I0226 03:17:26.773937  1117 leveldb.cpp:341] Persisting action (442 bytes) to 
leveldb took 469969ns
I0226 03:17:26.773957  1117 replica.cpp:712] Persisted action at 3
I0226 03:17:26.774605  1114 replica.cpp:691] Replica received learned notice 
for position 3 from @0.0.0.0:0
I0226 03:17:26.775961  1114 leveldb.cpp:341] Persisting action (444 bytes) to 
leveldb took 1.329435ms
I0226 03:17:26.775986  1114 replica.cpp:712] Persisted action at 3
I0226 03:17:26.776008  1114 replica.cpp:697] Replica learned APPEND action at 
position 3
I0226 03:17:26.777228  1115 slave.cpp:1321] Will retry registration in 
41.5608ms if necessary
I0226 03:17:26.777300  1112 registrar.cpp:484] Successfully updated the 
'registry' in 5.378048ms
I0226 03:17:26.777361  1114 master.cpp:4228] Ignoring register slave message 
from slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io) as admission 
is already in progress
I0226 03:17:26.777505  1113 log.cpp:702] Attempting to truncate the log to 3
I0226 03:17:26.777616  1111 coordinator.cpp:348] Coordinator attempting to 
write TRUNCATE action at position 4
I0226 03:17:26.778062  1114 slave.cpp:3482] Received ping from 
slave-observer(369)@172.30.2.151:51934
I0226 03:17:26.778139  1118 master.cpp:4308] Registered slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io) with cpu(*):2; mem(*):2048; disk(role1):2048; 
cpus(*):8; ports(*):[31000-32000]
I0226 03:17:26.778213  1113 replica.cpp:537] Replica received write request for 
position 4 from (13825)@172.30.2.151:51934
I0226 03:17:26.778291  1114 slave.cpp:971] Registered with master 
master@172.30.2.151:51934; given slave ID 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
I0226 03:17:26.778316  1114 fetcher.cpp:81] Clearing fetcher cache
I0226 03:17:26.778367  1116 hierarchical.cpp:473] Added slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io) with 
cpu(*):2; mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000] 
(allocated: )
I0226 03:17:26.778447  1117 status_update_manager.cpp:181] Resuming sending 
status updates
I0226 03:17:26.778617  1113 leveldb.cpp:341] Persisting action (16 bytes) to 
leveldb took 375414ns
I0226 03:17:26.778635  1113 replica.cpp:712] Persisted action at 4
I0226 03:17:26.778650  1114 slave.cpp:994] Checkpointing SlaveInfo to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/slave.info'
I0226 03:17:26.778900  1114 slave.cpp:1030] Forwarding total oversubscribed 
resources 
I0226 03:17:26.779109  1114 master.cpp:4649] Received update of slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io) with total oversubscribed resources 
I0226 03:17:26.779139  1112 replica.cpp:691] Replica received learned notice 
for position 4 from @0.0.0.0:0
I0226 03:17:26.779331  1116 hierarchical.cpp:1529] No inverse offers to send 
out!
I0226 03:17:26.779369  1116 hierarchical.cpp:1147] Performed allocation for 
slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 in 969593ns
I0226 03:17:26.779645  1113 master.cpp:5355] Sending 1 offers to framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934
I0226 03:17:26.779700  1116 hierarchical.cpp:531] Slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io) updated 
with oversubscribed resources  (total: cpu(*):2; mem(*):2048; disk(role1):2048; 
cpus(*):8; ports(*):[31000-32000], allocated: disk(role1):2048; cpu(*):2; 
mem(*):2048; cpus(*):8; ports(*):[31000-32000])
I0226 03:17:26.779819  1116 hierarchical.cpp:1434] No resources available to 
allocate!
I0226 03:17:26.779847  1116 hierarchical.cpp:1529] No inverse offers to send 
out!
I0226 03:17:26.779865  1116 hierarchical.cpp:1147] Performed allocation for 
slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 in 133437ns
I0226 03:17:26.780025  1118 sched.cpp:873] Scheduler::resourceOffers took 
102165ns
I0226 03:17:26.780372  1097 resources.cpp:576] Parsing resources as JSON 
failed: cpus:1;mem:64;
Trying semicolon-delimited string format instead
I0226 03:17:26.780882  1112 leveldb.cpp:341] Persisting action (18 bytes) to 
leveldb took 1.715066ms
I0226 03:17:26.780938  1112 leveldb.cpp:399] Deleting ~2 keys from leveldb took 
32370ns
I0226 03:17:26.780953  1112 replica.cpp:712] Persisted action at 4
I0226 03:17:26.780971  1112 replica.cpp:697] Replica learned TRUNCATE action at 
position 4
I0226 03:17:26.781693  1117 master.cpp:3138] Processing ACCEPT call for offers: 
[ 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-O0 ] on slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io) for framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934
I0226 03:17:26.781731  1117 master.cpp:2926] Authorizing principal 
'test-principal' to create volumes
I0226 03:17:26.781801  1117 master.cpp:2825] Authorizing framework principal 
'test-principal' to launch task 1 as user 'root'
I0226 03:17:26.782827  1114 master.cpp:3467] Applying CREATE operation for 
volumes disk(role1)[id1:path1]:64 from framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934 to slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io)
I0226 03:17:26.783136  1114 master.cpp:6589] Sending checkpointed resources 
disk(role1)[id1:path1]:64 to slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at 
slave(399)@172.30.2.151:51934 (ip-172-30-2-151.mesosphere.io)
I0226 03:17:26.783641  1111 slave.cpp:2341] Updated checkpointed resources from 
 to disk(role1)[id1:path1]:64
I0226 03:17:26.783911  1114 master.hpp:176] Adding task 1 with resources 
cpus(*):1; mem(*):64; disk(role1)[id1:path1]:64 on slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 (ip-172-30-2-151.mesosphere.io)
I0226 03:17:26.784056  1114 master.cpp:3623] Launching task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934 with 
resources cpus(*):1; mem(*):64; disk(role1)[id1:path1]:64 on slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io)
I0226 03:17:26.784397  1115 slave.cpp:1361] Got assigned task 1 for framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.784557  1115 slave.cpp:5287] Checkpointing FrameworkInfo to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/framework.info'
I0226 03:17:26.784739  1116 hierarchical.cpp:653] Updated allocation of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 on slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 from disk(role1):2048; cpu(*):2; 
mem(*):2048; cpus(*):8; ports(*):[31000-32000] to disk(role1):1984; cpu(*):2; 
mem(*):2048; cpus(*):8; ports(*):[31000-32000]; disk(role1)[id1:path1]:64
I0226 03:17:26.784848  1115 slave.cpp:5298] Checkpointing framework pid 
'scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934' to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/framework.pid'
I0226 03:17:26.785078  1115 resources.cpp:576] Parsing resources as JSON 
failed: cpus:0.1;mem:32
Trying semicolon-delimited string format instead
I0226 03:17:26.785322  1116 hierarchical.cpp:892] Recovered disk(role1):1984; 
cpu(*):2; mem(*):1984; cpus(*):7; ports(*):[31000-32000] (total: cpu(*):2; 
mem(*):2048; disk(role1):1984; cpus(*):8; ports(*):[31000-32000]; 
disk(role1)[id1:path1]:64, allocated: disk(role1)[id1:path1]:64; cpus(*):1; 
mem(*):64) on slave 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 from framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.785658  1115 slave.cpp:1480] Launching task 1 for framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.785719  1115 resources.cpp:576] Parsing resources as JSON 
failed: cpus:0.1;mem:32
Trying semicolon-delimited string format instead
I0226 03:17:26.786197  1115 paths.cpp:474] Trying to chown 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
 to user 'root'
I0226 03:17:26.791122  1115 slave.cpp:5739] Checkpointing ExecutorInfo to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/executor.info'
I0226 03:17:26.791543  1115 slave.cpp:5367] Launching executor 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 with resources cpus(*):0.1; mem(*):32 
in work directory 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
I0226 03:17:26.792325  1115 slave.cpp:5762] Checkpointing TaskInfo to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/tasks/1/task.info'
I0226 03:17:26.794337  1115 slave.cpp:1698] Queuing task '1' for executor '1' 
of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:26.794478  1115 slave.cpp:749] Successfully attached file 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
I0226 03:17:26.797106  1116 docker.cpp:1023] Starting container 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' for task '1' (and executor '1') of 
framework '5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000'
I0226 03:17:26.797462  1116 docker.cpp:1053] Running docker -H 
unix:///var/run/docker.sock inspect alpine:latest
I0226 03:17:26.910549  1111 docker.cpp:394] Docker pull alpine completed
I0226 03:17:26.910800  1111 docker.cpp:483] Changing the ownership of the 
persistent volume at 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/volumes/roles/role1/id1'
 with uid 0 and gid 0
I0226 03:17:26.915712  1111 docker.cpp:504] Mounting 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/volumes/roles/role1/id1'
 to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1'
 for persistent volume disk(role1)[id1:path1]:64 of container 
bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c
I0226 03:17:26.919000  1117 docker.cpp:576] Checkpointing pid 9568 to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/pids/forked.pid'
I0226 03:17:26.974776  1114 slave.cpp:2643] Got registration for executor '1' 
of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from 
executor(1)@172.30.2.151:46052
I0226 03:17:26.975217  1114 slave.cpp:2729] Checkpointing executor pid 
'executor(1)@172.30.2.151:46052' to 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/pids/libprocess.pid'
I0226 03:17:26.976177  1113 docker.cpp:1303] Ignoring updating container 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' with resources passed to update is 
identical to existing resources
I0226 03:17:26.976492  1115 slave.cpp:1863] Sending queued task '1' to executor 
'1' of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 at 
executor(1)@172.30.2.151:46052
I0226 03:17:27.691769  1111 slave.cpp:3002] Handling status update TASK_RUNNING 
(UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from executor(1)@172.30.2.151:46052
I0226 03:17:27.692291  1116 status_update_manager.cpp:320] Received status 
update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.692327  1116 status_update_manager.cpp:497] Creating 
StatusUpdate stream for task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.692773  1116 status_update_manager.cpp:824] Checkpointing UPDATE 
for status update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for 
task 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.700090  1116 status_update_manager.cpp:374] Forwarding update 
TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to the slave
I0226 03:17:27.700389  1113 slave.cpp:3400] Forwarding the update TASK_RUNNING 
(UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to master@172.30.2.151:51934
I0226 03:17:27.700606  1113 slave.cpp:3294] Status update manager successfully 
handled status update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) 
for task 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.700644  1113 slave.cpp:3310] Sending acknowledgement for status 
update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 to 
executor(1)@172.30.2.151:46052
I0226 03:17:27.700742  1117 master.cpp:4794] Status update TASK_RUNNING (UUID: 
9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 from slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io)
I0226 03:17:27.700775  1117 master.cpp:4842] Forwarding status update 
TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.700923  1117 master.cpp:6450] Updating the state of task 1 of 
framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (latest state: 
TASK_RUNNING, status update state: TASK_RUNNING)
I0226 03:17:27.701145  1118 sched.cpp:981] Scheduler::statusUpdate took 107222ns
I0226 03:17:27.701550  1112 master.cpp:3952] Processing ACKNOWLEDGE call 
9f75a4e5-9ff4-4ca9-8623-8b2574796229 for task 1 of framework 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000 (default) at 
scheduler-c59020d6-385e-48a3-8a10-9e5c3f1dbd92@172.30.2.151:51934 on slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0
I0226 03:17:27.701828  1114 status_update_manager.cpp:392] Received status 
update acknowledgement (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for task 1 
of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.701962  1114 status_update_manager.cpp:824] Checkpointing ACK 
for status update TASK_RUNNING (UUID: 9f75a4e5-9ff4-4ca9-8623-8b2574796229) for 
task 1 of framework 5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000
I0226 03:17:27.701987  1112 slave.cpp:668] Slave terminating
I0226 03:17:27.702256  1117 master.cpp:1174] Slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io) disconnected
I0226 03:17:27.702275  1117 master.cpp:2635] Disconnecting slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io)
I0226 03:17:27.702335  1117 master.cpp:2654] Deactivating slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 at slave(399)@172.30.2.151:51934 
(ip-172-30-2-151.mesosphere.io)
I0226 03:17:27.702492  1111 hierarchical.cpp:560] Slave 
5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0 deactivated
I0226 03:17:27.707713  1115 slave.cpp:193] Slave started on 
400)@172.30.2.151:51934
I0226 03:17:27.707739  1115 slave.cpp:194] Flags at startup: 
--appc_simple_discovery_uri_prefix="http://"; 
--appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5" 
--cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
--cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
--cgroups_root="mesos" --container_disk_watch_interval="15secs" 
--containerizers="mesos" 
--credential="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential"
 --default_role="*" --disk_watch_interval="1mins" --docker="docker" 
--docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io"; 
--docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock" 
--docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker" 
--enforce_container_disk_quota="false" --executor_registration_timeout="1mins" 
--executor_shutdown_grace_period="5secs" 
--fetcher_cache_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/fetch"
 --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
--gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
--hostname_lookup="true" --image_provisioner_backend="copy" 
--initialize_driver_logging="true" --isolation="posix/cpu,posix/mem" 
--launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src" --logbufsecs="0" 
--logging_level="INFO" --oversubscribed_resources_interval="15secs" 
--perf_duration="10secs" --perf_interval="1mins" 
--qos_correction_interval_min="0ns" --quiet="false" --recover="reconnect" 
--recovery_timeout="15mins" --registration_backoff_factor="10ms" 
--resources="cpu:2;mem:2048;disk(role1):2048" 
--revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox" 
--strict="true" --switch_user="true" --systemd_enable_support="true" 
--systemd_runtime_directory="/run/systemd/system" --version="false" 
--work_dir="/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP"
I0226 03:17:27.708133  1115 credentials.hpp:83] Loading credential for 
authentication from 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/credential'
I0226 03:17:27.708282  1115 slave.cpp:324] Slave using credential for: 
test-principal
I0226 03:17:27.708407  1115 resources.cpp:576] Parsing resources as JSON 
failed: cpu:2;mem:2048;disk(role1):2048
Trying semicolon-delimited string format instead
I0226 03:17:27.708874  1115 slave.cpp:464] Slave resources: cpu(*):2; 
mem(*):2048; disk(role1):2048; cpus(*):8; ports(*):[31000-32000]
I0226 03:17:27.708931  1115 slave.cpp:472] Slave attributes: [  ]
I0226 03:17:27.708941  1115 slave.cpp:477] Slave hostname: 
ip-172-30-2-151.mesosphere.io
I0226 03:17:27.710033  1113 state.cpp:58] Recovering state from 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta'
I0226 03:17:27.711252  1114 fetcher.cpp:81] Clearing fetcher cache
I0226 03:17:27.711447  1116 status_update_manager.cpp:200] Recovering status 
update manager
I0226 03:17:27.711727  1111 docker.cpp:726] Recovering Docker containers
I0226 03:17:27.711839  1111 docker.cpp:885] Running docker -H 
unix:///var/run/docker.sock ps -a
I0226 03:17:27.728170  1117 hierarchical.cpp:1434] No resources available to 
allocate!
I0226 03:17:27.728235  1117 hierarchical.cpp:1529] No inverse offers to send 
out!
I0226 03:17:27.728268  1117 hierarchical.cpp:1127] Performed allocation for 1 
slaves in 296715ns
I0226 03:17:27.817551  1113 docker.cpp:766] Running docker -H 
unix:///var/run/docker.sock inspect 
mesos-5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0.bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c
I0226 03:17:27.923014  1112 docker.cpp:932] Checking if Docker container named 
'/mesos-5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0.bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
 was started by Mesos
I0226 03:17:27.923071  1112 docker.cpp:942] Checking if Mesos container with ID 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c' has been orphaned
I0226 03:17:27.923122  1112 docker.cpp:678] Running docker -H 
unix:///var/run/docker.sock stop -t 0 
0a10ad8641f8e85227324a979817933322dc901706cb4430eab0bcaf979835d1
I0226 03:17:28.023885  1116 docker.cpp:727] Running docker -H 
unix:///var/run/docker.sock rm -v 
0a10ad8641f8e85227324a979817933322dc901706cb4430eab0bcaf979835d1

I0226 03:17:28.127876  1114 docker.cpp:912] Unmounting volume for container 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
../../3rdparty/libprocess/include/process/gmock.hpp:214: ERROR: this mock 
object (used in test 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should be 
deleted but never is. Its address is @0x5781dd8.
I0226 03:17:28.127957  1114 docker.cpp:912] Unmounting volume for container 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c'
../../src/tests/mesos.cpp:673: ERROR: this mock object (used in test 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should be 
deleted but never is. Its address is @0x5a03260.
../../src/tests/mesos.hpp:1357: ERROR: this mock object (used in test 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should be 
deleted but never is. Its address is @0x5b477c0.
Failed to perform recovery: Unable to unmount volumes for Docker container 
'bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c': Failed to unmount volume 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1':
 Failed to unmount 
'/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/slaves/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-S0/frameworks/5cc57c0e-f1ad-4107-893f-420ed1a1db1a-0000/executors/1/runs/bcc90102-163d-4ff6-a3fc-a1b2e3fc3b7c/path1':
 Invalid argument
../../src/tests/containerizer/docker_containerizer_tests.cpp:1650: ERROR: this 
mock object (used in test 
DockerContainerizerTest.ROOT_DOCKER_RecoverOrphanedPersistentVolumes) should be 
deleted but never is. Its address is @0x7ffe75a8d310.
To remedy this do as follows:
ERROR: 4 leaked mock objects found at program exit.
Step 1: rm -f 
/tmp/DockerContainerizerTest_ROOT_DOCKER_RecoverOrphanedPersistentVolumes_aJOesP/meta/slaves/latest
        This ensures slave doesn't recover old live executors.
Step 2: Restart the slave.
Process exited with code 1
{code}

There appear to be two problems:
1) The docker containerizer should not exit on failure to clean up orphans.  
The MesosContainerizer does not do this (see [MESOS-2367]).
2) Unmounting the orphan persistent volume fails for some reason.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to