[ 
https://issues.apache.org/jira/browse/MESOS-6336?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15570475#comment-15570475
 ] 

Vinod Kone commented on MESOS-6336:
-----------------------------------

Looking at the test code, one potential issue I can see is that we are calling 
a method (`unmocked__run`) of a libprocess actor (MockSlave) directly from the 
test code instead of dispatching to it. If two different threads are 
simultaneously accessing the `Slave` object thus causing seg fault.

Not sure why this is happening inside `Slave::finalize()` though because AFAICT 
`unmocked__run` should've been completely done (i.e., removed framework from 
the `frameworks` map) by the time `finalize()` gets called.

> SlaveTest.KillTaskGroupBetweenRunTaskParts is flaky
> ---------------------------------------------------
>
>                 Key: MESOS-6336
>                 URL: https://issues.apache.org/jira/browse/MESOS-6336
>             Project: Mesos
>          Issue Type: Bug
>          Components: slave
>            Reporter: Greg Mann
>              Labels: mesosphere
>
> The test {{SlaveTest.KillTaskGroupBetweenRunTaskParts}} sometimes segfaults 
> during the agent's {{finalize()}} method. This was observed on our internal 
> CI, on Fedora with libev, without SSL:
> {code}
> [ RUN      ] SlaveTest.KillTaskGroupBetweenRunTaskParts
> I1007 14:12:57.973811 28630 cluster.cpp:158] Creating default 'local' 
> authorizer
> I1007 14:12:57.982128 28630 leveldb.cpp:174] Opened db in 8.195028ms
> I1007 14:12:57.982599 28630 leveldb.cpp:181] Compacted db in 446238ns
> I1007 14:12:57.982616 28630 leveldb.cpp:196] Created db iterator in 3650ns
> I1007 14:12:57.982622 28630 leveldb.cpp:202] Seeked to beginning of db in 
> 451ns
> I1007 14:12:57.982627 28630 leveldb.cpp:271] Iterated through 0 keys in the 
> db in 352ns
> I1007 14:12:57.982638 28630 replica.cpp:776] Replica recovered with log 
> positions 0 -> 0 with 1 holes and 0 unlearned
> I1007 14:12:57.983024 28645 recover.cpp:451] Starting replica recovery
> I1007 14:12:57.983127 28651 recover.cpp:477] Replica is in EMPTY status
> I1007 14:12:57.983459 28644 replica.cpp:673] Replica in EMPTY status received 
> a broadcasted recover request from __req_res__(6234)@172.30.2.161:38776
> I1007 14:12:57.983543 28651 recover.cpp:197] Received a recover response from 
> a replica in EMPTY status
> I1007 14:12:57.983680 28650 recover.cpp:568] Updating replica status to 
> STARTING
> I1007 14:12:57.983990 28648 master.cpp:380] Master 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb 
> (ip-172-30-2-161.ec2.internal.mesosphere.io) started on 172.30.2.161:38776
> I1007 14:12:57.984007 28648 master.cpp:382] Flags at startup: --acls="" 
> --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" 
> --allocation_interval="1secs" --allocator="HierarchicalDRF" 
> --authenticate_agents="true" --authenticate_frameworks="true" 
> --authenticate_http_frameworks="true" --authenticate_http_readonly="true" 
> --authenticate_http_readwrite="true" --authenticators="crammd5" 
> --authorizers="local" --credentials="/tmp/rVbcaO/credentials" 
> --framework_sorter="drf" --help="false" --hostname_lookup="true" 
> --http_authenticators="basic" --http_framework_authenticators="basic" 
> --initialize_driver_logging="true" --log_auto_initialize="true" 
> --logbufsecs="0" --logging_level="INFO" --max_agent_ping_timeouts="5" 
> --max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" 
> --quiet="false" --recovery_agent_removal_limit="100%" 
> --registry="replicated_log" --registry_fetch_timeout="1mins" 
> --registry_gc_interval="15mins" --registry_max_agent_age="2weeks" 
> --registry_max_agent_count="102400" --registry_store_timeout="100secs" 
> --registry_strict="false" --root_submissions="true" --user_sorter="drf" 
> --version="false" --webui_dir="/usr/local/share/mesos/webui" 
> --work_dir="/tmp/rVbcaO/master" --zk_session_timeout="10secs"
> I1007 14:12:57.984127 28648 master.cpp:432] Master only allowing 
> authenticated frameworks to register
> I1007 14:12:57.984134 28648 master.cpp:446] Master only allowing 
> authenticated agents to register
> I1007 14:12:57.984139 28648 master.cpp:459] Master only allowing 
> authenticated HTTP frameworks to register
> I1007 14:12:57.984143 28648 credentials.hpp:37] Loading credentials for 
> authentication from '/tmp/rVbcaO/credentials'
> I1007 14:12:57.988487 28648 master.cpp:504] Using default 'crammd5' 
> authenticator
> I1007 14:12:57.988530 28648 http.cpp:883] Using default 'basic' HTTP 
> authenticator for realm 'mesos-master-readonly'
> I1007 14:12:57.988585 28648 http.cpp:883] Using default 'basic' HTTP 
> authenticator for realm 'mesos-master-readwrite'
> I1007 14:12:57.988648 28648 http.cpp:883] Using default 'basic' HTTP 
> authenticator for realm 'mesos-master-scheduler'
> I1007 14:12:57.988680 28648 master.cpp:584] Authorization enabled
> I1007 14:12:57.988757 28650 whitelist_watcher.cpp:77] No whitelist given
> I1007 14:12:57.988772 28646 hierarchical.cpp:149] Initialized hierarchical 
> allocator process
> I1007 14:12:57.988917 28651 leveldb.cpp:304] Persisting metadata (8 bytes) to 
> leveldb took 5.186917ms
> I1007 14:12:57.988934 28651 replica.cpp:320] Persisted replica status to 
> STARTING
> I1007 14:12:57.989045 28651 recover.cpp:477] Replica is in STARTING status
> I1007 14:12:57.989316 28648 master.cpp:2013] Elected as the leading master!
> I1007 14:12:57.989331 28648 master.cpp:1560] Recovering from registrar
> I1007 14:12:57.989408 28649 replica.cpp:673] Replica in STARTING status 
> received a broadcasted recover request from 
> __req_res__(6235)@172.30.2.161:38776
> I1007 14:12:57.989423 28648 registrar.cpp:329] Recovering registrar
> I1007 14:12:57.989792 28647 recover.cpp:197] Received a recover response from 
> a replica in STARTING status
> I1007 14:12:57.989956 28650 recover.cpp:568] Updating replica status to VOTING
> I1007 14:12:57.990365 28651 leveldb.cpp:304] Persisting metadata (8 bytes) to 
> leveldb took 326153ns
> I1007 14:12:57.990378 28651 replica.cpp:320] Persisted replica status to 
> VOTING
> I1007 14:12:57.990412 28651 recover.cpp:582] Successfully joined the Paxos 
> group
> I1007 14:12:57.990470 28651 recover.cpp:466] Recover process terminated
> I1007 14:12:57.990602 28647 log.cpp:553] Attempting to start the writer
> I1007 14:12:57.990994 28647 replica.cpp:493] Replica received implicit 
> promise request from __req_res__(6236)@172.30.2.161:38776 with proposal 1
> I1007 14:12:57.992638 28647 leveldb.cpp:304] Persisting metadata (8 bytes) to 
> leveldb took 1.62248ms
> I1007 14:12:57.992655 28647 replica.cpp:342] Persisted promised to 1
> I1007 14:12:57.992843 28646 coordinator.cpp:238] Coordinator attempting to 
> fill missing positions
> I1007 14:12:57.993249 28650 replica.cpp:388] Replica received explicit 
> promise request from __req_res__(6237)@172.30.2.161:38776 for position 0 with 
> proposal 2
> I1007 14:12:57.995498 28650 leveldb.cpp:341] Persisting action (8 bytes) to 
> leveldb took 2.225608ms
> I1007 14:12:57.995515 28650 replica.cpp:708] Persisted action NOP at position > 0
> I1007 14:12:57.996105 28649 replica.cpp:537] Replica received write request 
> for position 0 from __req_res__(6238)@172.30.2.161:38776
> I1007 14:12:57.996130 28649 leveldb.cpp:436] Reading position from leveldb 
> took 8903ns
> I1007 14:12:57.997609 28649 leveldb.cpp:341] Persisting action (14 bytes) to 
> leveldb took 1.465755ms
> I1007 14:12:57.997627 28649 replica.cpp:708] Persisted action NOP at position > 0
> I1007 14:12:57.997859 28646 replica.cpp:691] Replica received learned notice 
> for position 0 from @0.0.0.0:0
> I1007 14:12:57.999869 28646 leveldb.cpp:341] Persisting action (16 bytes) to 
> leveldb took 1.985345ms
> I1007 14:12:57.999886 28646 replica.cpp:708] Persisted action NOP at position > 0
> I1007 14:12:58.000043 28644 log.cpp:569] Writer started with ending position 0
> I1007 14:12:58.000308 28650 leveldb.cpp:436] Reading position from leveldb 
> took 8280ns
> I1007 14:12:58.000512 28650 registrar.cpp:362] Successfully fetched the 
> registry (0B) in 11.06816ms
> I1007 14:12:58.000562 28650 registrar.cpp:461] Applied 1 operations in 
> 3050ns; attempting to update the registry
> I1007 14:12:58.000722 28646 log.cpp:577] Attempting to append 235 bytes to 
> the log
> I1007 14:12:58.000799 28651 coordinator.cpp:348] Coordinator attempting to 
> write APPEND action at position 1
> I1007 14:12:58.001106 28648 replica.cpp:537] Replica received write request 
> for position 1 from __req_res__(6239)@172.30.2.161:38776
> I1007 14:12:58.002013 28648 leveldb.cpp:341] Persisting action (254 bytes) to 
> leveldb took 884026ns
> I1007 14:12:58.002027 28648 replica.cpp:708] Persisted action APPEND at 
> position 1
> I1007 14:12:58.002336 28650 replica.cpp:691] Replica received learned notice 
> for position 1 from @0.0.0.0:0
> I1007 14:12:58.004251 28650 leveldb.cpp:341] Persisting action (256 bytes) to 
> leveldb took 1.897221ms
> I1007 14:12:58.004268 28650 replica.cpp:708] Persisted action APPEND at 
> position 1
> I1007 14:12:58.004483 28645 registrar.cpp:506] Successfully updated the 
> registry in 3.902208ms
> I1007 14:12:58.004537 28646 log.cpp:596] Attempting to truncate the log to 1
> I1007 14:12:58.004566 28644 coordinator.cpp:348] Coordinator attempting to 
> write TRUNCATE action at position 2
> I1007 14:12:58.004580 28645 registrar.cpp:392] Successfully recovered 
> registrar
> I1007 14:12:58.004871 28650 master.cpp:1676] Recovered 0 agents from the 
> registry (196B); allowing 10mins for agents to re-register
> I1007 14:12:58.004896 28648 replica.cpp:537] Replica received write request 
> for position 2 from __req_res__(6240)@172.30.2.161:38776
> I1007 14:12:58.004914 28649 hierarchical.cpp:176] Skipping recovery of 
> hierarchical allocator: nothing to recover
> I1007 14:12:58.006487 28648 leveldb.cpp:341] Persisting action (16 bytes) to 
> leveldb took 1.566923ms
> I1007 14:12:58.006503 28648 replica.cpp:708] Persisted action TRUNCATE at 
> position 2
> I1007 14:12:58.006700 28651 replica.cpp:691] Replica received learned notice 
> for position 2 from @0.0.0.0:0
> I1007 14:12:58.008687 28651 leveldb.cpp:341] Persisting action (18 bytes) to 
> leveldb took 1.966404ms
> I1007 14:12:58.008713 28651 leveldb.cpp:399] Deleting ~1 keys from leveldb 
> took 10225ns
> I1007 14:12:58.008721 28651 replica.cpp:708] Persisted action TRUNCATE at 
> position 2
> I1007 14:12:58.010568 28645 slave.cpp:208] Mesos agent started on 
> @172.30.2.161:38776
> I1007 14:12:58.010679 28630 scheduler.cpp:176] Version: 1.1.0
> I1007 14:12:58.010578 28645 slave.cpp:209] Flags at startup: --acls="" 
> --appc_simple_discovery_uri_prefix="http://"; 
> --appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="true" 
> --authenticate_http_readwrite="true" --authenticatee="crammd5" 
> --authentication_backoff_factor="1secs" --authorizer="local" 
> --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" 
> --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" 
> --cgroups_root="mesos" --container_disk_watch_interval="15secs" 
> --containerizers="mesos" 
> --credential="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/credential"
>  --default_role="*" --disk_watch_interval="1mins" --docker="docker" 
> --docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io"; 
> --docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock" 
> --docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker" 
> --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" 
> --enforce_container_disk_quota="false" 
> --executor_registration_timeout="1mins" 
> --executor_shutdown_grace_period="5secs" 
> --fetcher_cache_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/fetch"
>  --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" 
> --gc_disk_headroom="0.1" --hadoop_home="" --help="false" 
> --hostname_lookup="true" --http_authenticators="basic" 
> --http_command_executor="false" 
> --http_credentials="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/http_credentials"
>  --image_provisioner_backend="copy" --initialize_driver_logging="true" 
> --isolation="posix/cpu,posix/mem" --launcher="linux" 
> --launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src" 
> --logbufsecs="0" --logging_level="INFO" 
> --oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
> --perf_interval="1mins" --qos_correction_interval_min="0ns" --quiet="false" 
> --recover="reconnect" --recovery_timeout="15mins" 
> --registration_backoff_factor="10ms" 
> --resources="cpus:2;gpus:0;mem:1024;disk:1024;ports:[31000-32000]" 
> --revocable_cpu_low_priority="true" 
> --runtime_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5"
>  --sandbox_directory="/mnt/mesos/sandbox" --strict="true" 
> --switch_user="true" --systemd_enable_support="true" 
> --systemd_runtime_directory="/run/systemd/system" --version="false" 
> --work_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW"
>  --xfs_project_range="[5000-10000]"
> I1007 14:12:58.010781 28645 credentials.hpp:86] Loading credential for 
> authentication from 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/credential'
> I1007 14:12:58.010834 28645 slave.cpp:346] Agent using credential for: 
> test-principal
> I1007 14:12:58.010850 28645 credentials.hpp:37] Loading credentials for 
> authentication from 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/http_credentials'
> I1007 14:12:58.010880 28644 scheduler.cpp:465] New master detected at 
> master@172.30.2.161:38776
> I1007 14:12:58.010898 28644 scheduler.cpp:474] Waiting for 0ns before 
> initiating a re-(connection) attempt with the master
> I1007 14:12:58.010920 28645 http.cpp:883] Using default 'basic' HTTP 
> authenticator for realm 'mesos-agent-readonly'
> I1007 14:12:58.010948 28645 http.cpp:883] Using default 'basic' HTTP 
> authenticator for realm 'mesos-agent-readwrite'
> I1007 14:12:58.011189 28645 slave.cpp:533] Agent resources: cpus(*):2; 
> mem(*):1024; disk(*):1024; ports(*):[31000-32000]
> I1007 14:12:58.011216 28645 slave.cpp:541] Agent attributes: [  ]
> I1007 14:12:58.011221 28645 slave.cpp:546] Agent hostname: 
> ip-172-30-2-161.ec2.internal.mesosphere.io
> I1007 14:12:58.011471 28651 state.cpp:57] Recovering state from 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/meta'
> I1007 14:12:58.011592 28644 status_update_manager.cpp:203] Recovering status 
> update manager
> I1007 14:12:58.011668 28644 slave.cpp:5256] Finished recovery
> I1007 14:12:58.011844 28644 slave.cpp:5428] Querying resource estimator for 
> oversubscribable resources
> I1007 14:12:58.011942 28644 slave.cpp:915] New master detected at 
> master@172.30.2.161:38776
> I1007 14:12:58.011958 28651 status_update_manager.cpp:177] Pausing sending 
> status updates
> I1007 14:12:58.011960 28644 slave.cpp:974] Authenticating with master 
> master@172.30.2.161:38776
> I1007 14:12:58.011991 28644 slave.cpp:985] Using default CRAM-MD5 
> authenticatee
> I1007 14:12:58.012025 28644 slave.cpp:947] Detecting new master
> I1007 14:12:58.012051 28649 authenticatee.cpp:121] Creating new client SASL 
> connection
> I1007 14:12:58.012137 28651 master.cpp:6634] Authenticating 
> (401)@172.30.2.161:38776
> I1007 14:12:58.012226 28647 authenticator.cpp:414] Starting authentication 
> session for crammd5-authenticatee(991)@172.30.2.161:38776
> I1007 14:12:58.012310 28651 authenticator.cpp:98] Creating new server SASL 
> connection
> I1007 14:12:58.012409 28651 authenticatee.cpp:213] Received SASL 
> authentication mechanisms: CRAM-MD5
> I1007 14:12:58.012428 28651 authenticatee.cpp:239] Attempting to authenticate 
> with mechanism 'CRAM-MD5'
> I1007 14:12:58.012471 28651 authenticator.cpp:204] Received SASL 
> authentication start
> I1007 14:12:58.012507 28651 authenticator.cpp:326] Authentication requires 
> more steps
> I1007 14:12:58.012538 28648 scheduler.cpp:353] Connected with the master at 
> http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.012552 28647 authenticatee.cpp:259] Received SASL 
> authentication step
> I1007 14:12:58.012673 28651 authenticator.cpp:232] Received SASL 
> authentication step
> I1007 14:12:58.012693 28651 auxprop.cpp:109] Request to lookup properties for 
> user: 'test-principal' realm: 'ip-172-30-2-161.ec2.internal' server FQDN: 
> 'ip-172-30-2-161.ec2.internal' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false 
> I1007 14:12:58.012701 28651 auxprop.cpp:181] Looking up auxiliary property 
> '*userPassword'
> I1007 14:12:58.012712 28651 auxprop.cpp:181] Looking up auxiliary property 
> '*cmusaslsecretCRAM-MD5'
> I1007 14:12:58.012722 28651 auxprop.cpp:109] Request to lookup properties for 
> user: 'test-principal' realm: 'ip-172-30-2-161.ec2.internal' server FQDN: 
> 'ip-172-30-2-161.ec2.internal' SASL_AUXPROP_VERIFY_AGAINST_HASH: false 
> SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true 
> I1007 14:12:58.012729 28651 auxprop.cpp:131] Skipping auxiliary property 
> '*userPassword' since SASL_AUXPROP_AUTHZID == true
> I1007 14:12:58.012737 28651 auxprop.cpp:131] Skipping auxiliary property 
> '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true
> I1007 14:12:58.012749 28651 authenticator.cpp:318] Authentication success
> I1007 14:12:58.012830 28648 authenticatee.cpp:299] Authentication success
> I1007 14:12:58.012850 28645 master.cpp:6664] Successfully authenticated 
> principal 'test-principal' at (401)@172.30.2.161:38776
> I1007 14:12:58.012876 28646 authenticator.cpp:432] Authentication session 
> cleanup for crammd5-authenticatee(991)@172.30.2.161:38776
> I1007 14:12:58.012972 28648 slave.cpp:1069] Successfully authenticated with 
> master master@172.30.2.161:38776
> I1007 14:12:58.013017 28651 scheduler.cpp:235] Sending SUBSCRIBE call to 
> http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.013093 28648 slave.cpp:1475] Will retry registration in 
> 4.168895ms if necessary
> I1007 14:12:58.013185 28649 master.cpp:5074] Registering agent at 
> (401)@172.30.2.161:38776 (ip-172-30-2-161.ec2.internal.mesosphere.io) with id 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0
> I1007 14:12:58.013326 28649 process.cpp:3377] Handling HTTP event for process 
> 'master' with path: '/master/api/v1/scheduler'
> I1007 14:12:58.013386 28644 registrar.cpp:461] Applied 1 operations in 
> 14823ns; attempting to update the registry
> I1007 14:12:58.013556 28651 log.cpp:577] Attempting to append 434 bytes to 
> the log
> I1007 14:12:58.013615 28651 coordinator.cpp:348] Coordinator attempting to 
> write APPEND action at position 3
> I1007 14:12:58.013741 28650 http.cpp:382] HTTP POST for 
> /master/api/v1/scheduler from 172.30.2.161:48518
> I1007 14:12:58.013804 28650 master.cpp:2309] Received subscription request 
> for HTTP framework 'default'
> I1007 14:12:58.013823 28650 master.cpp:2049] Authorizing framework principal 
> 'test-principal' to receive offers for role '*'
> I1007 14:12:58.013964 28648 master.cpp:2407] Subscribing framework 'default' 
> with checkpointing disabled and capabilities [  ]
> I1007 14:12:58.013985 28649 replica.cpp:537] Replica received write request 
> for position 3 from __req_res__(6241)@172.30.2.161:38776
> I1007 14:12:58.014142 28648 master.hpp:2162] Sending heartbeat to 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.014166 28644 hierarchical.cpp:275] Added framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.014189 28644 hierarchical.cpp:1694] No allocations performed
> I1007 14:12:58.014200 28644 hierarchical.cpp:1789] No inverse offers to send 
> out!
> I1007 14:12:58.014214 28644 hierarchical.cpp:1286] Performed allocation for 0 
> agents in 31808ns
> I1007 14:12:58.014370 28645 scheduler.cpp:671] Enqueuing event SUBSCRIBED 
> received from http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.014456 28649 leveldb.cpp:341] Persisting action (453 bytes) to 
> leveldb took 448489ns
> I1007 14:12:58.014470 28649 replica.cpp:708] Persisted action APPEND at 
> position 3
> I1007 14:12:58.014523 28648 scheduler.cpp:671] Enqueuing event HEARTBEAT 
> received from http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.014883 28645 replica.cpp:691] Replica received learned notice 
> for position 3 from @0.0.0.0:0
> I1007 14:12:58.016865 28645 leveldb.cpp:341] Persisting action (455 bytes) to 
> leveldb took 1.963645ms
> I1007 14:12:58.016880 28645 replica.cpp:708] Persisted action APPEND at 
> position 3
> I1007 14:12:58.017127 28644 registrar.cpp:506] Successfully updated the 
> registry in 3.705088ms
> I1007 14:12:58.017184 28648 log.cpp:596] Attempting to truncate the log to 3
> I1007 14:12:58.017242 28651 coordinator.cpp:348] Coordinator attempting to 
> write TRUNCATE action at position 4
> I1007 14:12:58.017367 28645 master.cpp:5145] Registered agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io) with cpus(*):2; mem(*):1024; 
> disk(*):1024; ports(*):[31000-32000]
> I1007 14:12:58.017423 28644 slave.cpp:4108] Received ping from 
> slave-observer(454)@172.30.2.161:38776
> I1007 14:12:58.017436 28650 hierarchical.cpp:485] Added agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 
> (ip-172-30-2-161.ec2.internal.mesosphere.io) with cpus(*):2; mem(*):1024; 
> disk(*):1024; ports(*):[31000-32000] (allocated: {})
> I1007 14:12:58.017483 28644 slave.cpp:1115] Registered with master 
> master@172.30.2.161:38776; given agent ID 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0
> I1007 14:12:58.017501 28644 fetcher.cpp:86] Clearing fetcher cache
> I1007 14:12:58.017575 28651 status_update_manager.cpp:184] Resuming sending 
> status updates
> I1007 14:12:58.017680 28650 hierarchical.cpp:1789] No inverse offers to send 
> out!
> I1007 14:12:58.017709 28650 hierarchical.cpp:1309] Performed allocation for 
> agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 in 246743ns
> I1007 14:12:58.017750 28644 slave.cpp:1138] Checkpointing SlaveInfo to 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/meta/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/slave.info'
> I1007 14:12:58.017774 28645 master.cpp:6463] Sending 1 offers to framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default)
> I1007 14:12:58.017982 28644 replica.cpp:537] Replica received write request 
> for position 4 from __req_res__(6242)@172.30.2.161:38776
> I1007 14:12:58.018200 28645 scheduler.cpp:671] Enqueuing event OFFERS 
> received from http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.018880 28646 scheduler.cpp:235] Sending ACCEPT call to 
> http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.019086 28644 leveldb.cpp:341] Persisting action (16 bytes) to 
> leveldb took 1.082078ms
> I1007 14:12:58.019104 28644 replica.cpp:708] Persisted action TRUNCATE at 
> position 4
> I1007 14:12:58.019212 28645 process.cpp:3377] Handling HTTP event for process 
> 'master' with path: '/master/api/v1/scheduler'
> I1007 14:12:58.019390 28645 replica.cpp:691] Replica received learned notice 
> for position 4 from @0.0.0.0:0
> I1007 14:12:58.019577 28647 http.cpp:382] HTTP POST for 
> /master/api/v1/scheduler from 172.30.2.161:48516
> I1007 14:12:58.019711 28647 master.cpp:3521] Processing ACCEPT call for 
> offers: [ 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-O0 ] on agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io) for framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default)
> I1007 14:12:58.019737 28647 master.cpp:3143] Authorizing framework principal 
> 'test-principal' to launch task ed09b29b-d149-46c1-a31d-f8fe3fdf214e
> I1007 14:12:58.019785 28647 master.cpp:3143] Authorizing framework principal 
> 'test-principal' to launch task 9ce218ba-84b7-4ea4-a560-b30472f4a22c
> I1007 14:12:58.020279 28647 master.cpp:8214] Adding task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e with resources cpus(*):0.1; mem(*):32; 
> disk(*):32 on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 
> (ip-172-30-2-161.ec2.internal.mesosphere.io)
> I1007 14:12:58.020342 28647 master.cpp:8214] Adding task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c with resources cpus(*):0.1; mem(*):32; 
> disk(*):32 on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 
> (ip-172-30-2-161.ec2.internal.mesosphere.io)
> I1007 14:12:58.020375 28647 master.cpp:4358] Launching task group { 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c, ed09b29b-d149-46c1-a31d-f8fe3fdf214e } 
> of framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default) with 
> resources cpus(*):0.2; mem(*):64; disk(*):64 on agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io)
> I1007 14:12:58.020480 28648 slave.cpp:1539] Got assigned task group 
> containing tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] for framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.020515 28650 hierarchical.cpp:1018] Recovered cpus(*):1.8; 
> mem(*):960; disk(*):960; ports(*):[31000-32000] (total: cpus(*):2; 
> mem(*):1024; disk(*):1024; ports(*):[31000-32000], allocated: cpus(*):0.2; 
> mem(*):64; disk(*):64) on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from 
> framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.020541 28650 hierarchical.cpp:1055] Framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 filtered agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 for 5secs
> I1007 14:12:58.020830 28644 scheduler.cpp:235] Sending KILL call to 
> http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.021075 28651 process.cpp:3377] Handling HTTP event for process 
> 'master' with path: '/master/api/v1/scheduler'
> I1007 14:12:58.021347 28648 http.cpp:382] HTTP POST for 
> /master/api/v1/scheduler from 172.30.2.161:48516
> I1007 14:12:58.021399 28648 master.cpp:4682] Telling agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io) to kill task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default)
> I1007 14:12:58.021473 28646 slave.cpp:2283] Asked to kill task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> W1007 14:12:58.021491 28646 slave.cpp:2324] Killing task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 before it was launched
> I1007 14:12:58.021483 28645 leveldb.cpp:341] Persisting action (18 bytes) to 
> leveldb took 2.0668ms
> I1007 14:12:58.021524 28645 leveldb.cpp:399] Deleting ~2 keys from leveldb 
> took 17427ns
> I1007 14:12:58.021539 28645 replica.cpp:708] Persisted action TRUNCATE at 
> position 4
> I1007 14:12:58.021720 28630 slave.cpp:1696] Launching task group containing 
> tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] for framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> W1007 14:12:58.021749 28630 slave.cpp:1736] Ignoring running task group 
> containing tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 because it has been killed in the 
> meantime
> I1007 14:12:58.021777 28630 slave.cpp:3609] Handling status update 
> TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from @0.0.0.0:0
> W1007 14:12:58.021800 28630 slave.cpp:3705] Could not find the executor for 
> status update TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for 
> task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.021888 28630 slave.cpp:3609] Handling status update 
> TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from @0.0.0.0:0
> I1007 14:12:58.021890 28647 status_update_manager.cpp:323] Received status 
> update TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.021919 28647 status_update_manager.cpp:500] Creating 
> StatusUpdate stream for task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of 
> framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> W1007 14:12:58.021913 28630 slave.cpp:3705] Could not find the executor for 
> status update TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for 
> task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.021981 28630 slave.cpp:4709] Cleaning up framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.021986 28647 status_update_manager.cpp:377] Forwarding update 
> TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to the agent
> E1007 14:12:58.022033 28630 slave.cpp:5408] Failed to find the mtime of 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/frameworks/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000':
>  Error invoking stat for 
> '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/frameworks/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000':
>  No such file or directory
> I1007 14:12:58.022058 28651 slave.cpp:4026] Forwarding the update TASK_KILLED 
> (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to master@172.30.2.161:38776
> I1007 14:12:58.022063 28647 status_update_manager.cpp:323] Received status 
> update TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022099 28647 status_update_manager.cpp:500] Creating 
> StatusUpdate stream for task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of 
> framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022140 28651 slave.cpp:3920] Status update manager 
> successfully handled status update TASK_KILLED (UUID: 
> b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022178 28647 status_update_manager.cpp:377] Forwarding update 
> TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to the agent
> I1007 14:12:58.022231 28646 master.cpp:5672] Status update TASK_KILLED (UUID: 
> b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io)
> I1007 14:12:58.022239 28651 slave.cpp:4026] Forwarding the update TASK_KILLED 
> (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to master@172.30.2.161:38776
> I1007 14:12:58.022256 28646 master.cpp:5734] Forwarding status update 
> TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022279 28647 status_update_manager.cpp:285] Closing status 
> update streams for framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022294 28647 status_update_manager.cpp:531] Cleaning up status 
> update stream for task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022301 28651 slave.cpp:3920] Status update manager 
> successfully handled status update TASK_KILLED (UUID: 
> a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022330 28647 status_update_manager.cpp:531] Cleaning up status 
> update stream for task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022347 28646 master.cpp:7592] Updating the state of task 
> ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (latest state: TASK_KILLED, status 
> update state: TASK_KILLED)
> I1007 14:12:58.022449 28646 master.cpp:5672] Status update TASK_KILLED (UUID: 
> a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 
> (ip-172-30-2-161.ec2.internal.mesosphere.io)
> I1007 14:12:58.022464 28646 master.cpp:5734] Forwarding status update 
> TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022461 28644 hierarchical.cpp:1018] Recovered cpus(*):0.1; 
> mem(*):32; disk(*):32 (total: cpus(*):2; mem(*):1024; disk(*):1024; 
> ports(*):[31000-32000], allocated: cpus(*):0.1; mem(*):32; disk(*):32) on 
> agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022600 28650 scheduler.cpp:671] Enqueuing event UPDATE 
> received from http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.022609 28646 master.cpp:7592] Updating the state of task 
> 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (latest state: TASK_KILLED, status 
> update state: TASK_KILLED)
> I1007 14:12:58.022825 28644 hierarchical.cpp:1018] Recovered cpus(*):0.1; 
> mem(*):32; disk(*):32 (total: cpus(*):2; mem(*):1024; disk(*):1024; 
> ports(*):[31000-32000], allocated: {}) on agent 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from framework 
> 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000
> I1007 14:12:58.022981 28645 scheduler.cpp:671] Enqueuing event UPDATE 
> received from http://172.30.2.161:38776/master/api/v1/scheduler
> I1007 14:12:58.023365 28651 slave.cpp:787] Agent terminating
> *** Aborted at 1475849578 (unix time) try "date -d @1475849578" if you are 
> using GNU date ***
> PC: @     0x7fdbf26dc95f mesos::internal::slave::Slave::finalize()
> *** SIGSEGV (@0x88) received by PID 28630 (TID 0x7fdbe5910700) from PID 136; 
> stack trace: ***
>     @     0x7fdbdef6fbb2 (unknown)
>     @     0x7fdbdef74009 (unknown)
>     @     0x7fdbdef67f08 (unknown)
>     @     0x7fdbf14bc9f0 (unknown)
>     @     0x7fdbf26dc95f mesos::internal::slave::Slave::finalize()
>     @     0x7fdbf2ec461b process::ProcessManager::resume()
>     @     0x7fdbf2ec7cb7 
> _ZNSt6thread5_ImplISt12_Bind_simpleIFZN7process14ProcessManager12init_threadsEvEUt_vEEE6_M_runEv
>     @     0x7fdbf0dd6f20 (unknown)
>     @     0x7fdbf14b360a start_thread
>     @     0x7fdbf054659d __clone
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to