[ https://issues.apache.org/jira/browse/MESOS-6336?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15606172#comment-15606172 ]
Abhishek Dasgupta commented on MESOS-6336: ------------------------------------------ RR: https://reviews.apache.org/r/53173/ > SlaveTest.KillTaskGroupBetweenRunTaskParts is flaky > --------------------------------------------------- > > Key: MESOS-6336 > URL: https://issues.apache.org/jira/browse/MESOS-6336 > Project: Mesos > Issue Type: Bug > Components: slave > Reporter: Greg Mann > Assignee: Abhishek Dasgupta > Labels: mesosphere > > The test {{SlaveTest.KillTaskGroupBetweenRunTaskParts}} sometimes segfaults > during the agent's {{finalize()}} method. This was observed on our internal > CI, on Fedora with libev, without SSL: > {code} > [ RUN ] SlaveTest.KillTaskGroupBetweenRunTaskParts > I1007 14:12:57.973811 28630 cluster.cpp:158] Creating default 'local' > authorizer > I1007 14:12:57.982128 28630 leveldb.cpp:174] Opened db in 8.195028ms > I1007 14:12:57.982599 28630 leveldb.cpp:181] Compacted db in 446238ns > I1007 14:12:57.982616 28630 leveldb.cpp:196] Created db iterator in 3650ns > I1007 14:12:57.982622 28630 leveldb.cpp:202] Seeked to beginning of db in > 451ns > I1007 14:12:57.982627 28630 leveldb.cpp:271] Iterated through 0 keys in the > db in 352ns > I1007 14:12:57.982638 28630 replica.cpp:776] Replica recovered with log > positions 0 -> 0 with 1 holes and 0 unlearned > I1007 14:12:57.983024 28645 recover.cpp:451] Starting replica recovery > I1007 14:12:57.983127 28651 recover.cpp:477] Replica is in EMPTY status > I1007 14:12:57.983459 28644 replica.cpp:673] Replica in EMPTY status received > a broadcasted recover request from __req_res__(6234)@172.30.2.161:38776 > I1007 14:12:57.983543 28651 recover.cpp:197] Received a recover response from > a replica in EMPTY status > I1007 14:12:57.983680 28650 recover.cpp:568] Updating replica status to > STARTING > I1007 14:12:57.983990 28648 master.cpp:380] Master > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb > (ip-172-30-2-161.ec2.internal.mesosphere.io) started on 172.30.2.161:38776 > I1007 14:12:57.984007 28648 master.cpp:382] Flags at startup: --acls="" > --agent_ping_timeout="15secs" --agent_reregister_timeout="10mins" > --allocation_interval="1secs" --allocator="HierarchicalDRF" > --authenticate_agents="true" --authenticate_frameworks="true" > --authenticate_http_frameworks="true" --authenticate_http_readonly="true" > --authenticate_http_readwrite="true" --authenticators="crammd5" > --authorizers="local" --credentials="/tmp/rVbcaO/credentials" > --framework_sorter="drf" --help="false" --hostname_lookup="true" > --http_authenticators="basic" --http_framework_authenticators="basic" > --initialize_driver_logging="true" --log_auto_initialize="true" > --logbufsecs="0" --logging_level="INFO" --max_agent_ping_timeouts="5" > --max_completed_frameworks="50" --max_completed_tasks_per_framework="1000" > --quiet="false" --recovery_agent_removal_limit="100%" > --registry="replicated_log" --registry_fetch_timeout="1mins" > --registry_gc_interval="15mins" --registry_max_agent_age="2weeks" > --registry_max_agent_count="102400" --registry_store_timeout="100secs" > --registry_strict="false" --root_submissions="true" --user_sorter="drf" > --version="false" --webui_dir="/usr/local/share/mesos/webui" > --work_dir="/tmp/rVbcaO/master" --zk_session_timeout="10secs" > I1007 14:12:57.984127 28648 master.cpp:432] Master only allowing > authenticated frameworks to register > I1007 14:12:57.984134 28648 master.cpp:446] Master only allowing > authenticated agents to register > I1007 14:12:57.984139 28648 master.cpp:459] Master only allowing > authenticated HTTP frameworks to register > I1007 14:12:57.984143 28648 credentials.hpp:37] Loading credentials for > authentication from '/tmp/rVbcaO/credentials' > I1007 14:12:57.988487 28648 master.cpp:504] Using default 'crammd5' > authenticator > I1007 14:12:57.988530 28648 http.cpp:883] Using default 'basic' HTTP > authenticator for realm 'mesos-master-readonly' > I1007 14:12:57.988585 28648 http.cpp:883] Using default 'basic' HTTP > authenticator for realm 'mesos-master-readwrite' > I1007 14:12:57.988648 28648 http.cpp:883] Using default 'basic' HTTP > authenticator for realm 'mesos-master-scheduler' > I1007 14:12:57.988680 28648 master.cpp:584] Authorization enabled > I1007 14:12:57.988757 28650 whitelist_watcher.cpp:77] No whitelist given > I1007 14:12:57.988772 28646 hierarchical.cpp:149] Initialized hierarchical > allocator process > I1007 14:12:57.988917 28651 leveldb.cpp:304] Persisting metadata (8 bytes) to > leveldb took 5.186917ms > I1007 14:12:57.988934 28651 replica.cpp:320] Persisted replica status to > STARTING > I1007 14:12:57.989045 28651 recover.cpp:477] Replica is in STARTING status > I1007 14:12:57.989316 28648 master.cpp:2013] Elected as the leading master! > I1007 14:12:57.989331 28648 master.cpp:1560] Recovering from registrar > I1007 14:12:57.989408 28649 replica.cpp:673] Replica in STARTING status > received a broadcasted recover request from > __req_res__(6235)@172.30.2.161:38776 > I1007 14:12:57.989423 28648 registrar.cpp:329] Recovering registrar > I1007 14:12:57.989792 28647 recover.cpp:197] Received a recover response from > a replica in STARTING status > I1007 14:12:57.989956 28650 recover.cpp:568] Updating replica status to VOTING > I1007 14:12:57.990365 28651 leveldb.cpp:304] Persisting metadata (8 bytes) to > leveldb took 326153ns > I1007 14:12:57.990378 28651 replica.cpp:320] Persisted replica status to > VOTING > I1007 14:12:57.990412 28651 recover.cpp:582] Successfully joined the Paxos > group > I1007 14:12:57.990470 28651 recover.cpp:466] Recover process terminated > I1007 14:12:57.990602 28647 log.cpp:553] Attempting to start the writer > I1007 14:12:57.990994 28647 replica.cpp:493] Replica received implicit > promise request from __req_res__(6236)@172.30.2.161:38776 with proposal 1 > I1007 14:12:57.992638 28647 leveldb.cpp:304] Persisting metadata (8 bytes) to > leveldb took 1.62248ms > I1007 14:12:57.992655 28647 replica.cpp:342] Persisted promised to 1 > I1007 14:12:57.992843 28646 coordinator.cpp:238] Coordinator attempting to > fill missing positions > I1007 14:12:57.993249 28650 replica.cpp:388] Replica received explicit > promise request from __req_res__(6237)@172.30.2.161:38776 for position 0 with > proposal 2 > I1007 14:12:57.995498 28650 leveldb.cpp:341] Persisting action (8 bytes) to > leveldb took 2.225608ms > I1007 14:12:57.995515 28650 replica.cpp:708] Persisted action NOP at position > 0 > I1007 14:12:57.996105 28649 replica.cpp:537] Replica received write request > for position 0 from __req_res__(6238)@172.30.2.161:38776 > I1007 14:12:57.996130 28649 leveldb.cpp:436] Reading position from leveldb > took 8903ns > I1007 14:12:57.997609 28649 leveldb.cpp:341] Persisting action (14 bytes) to > leveldb took 1.465755ms > I1007 14:12:57.997627 28649 replica.cpp:708] Persisted action NOP at position > 0 > I1007 14:12:57.997859 28646 replica.cpp:691] Replica received learned notice > for position 0 from @0.0.0.0:0 > I1007 14:12:57.999869 28646 leveldb.cpp:341] Persisting action (16 bytes) to > leveldb took 1.985345ms > I1007 14:12:57.999886 28646 replica.cpp:708] Persisted action NOP at position > 0 > I1007 14:12:58.000043 28644 log.cpp:569] Writer started with ending position 0 > I1007 14:12:58.000308 28650 leveldb.cpp:436] Reading position from leveldb > took 8280ns > I1007 14:12:58.000512 28650 registrar.cpp:362] Successfully fetched the > registry (0B) in 11.06816ms > I1007 14:12:58.000562 28650 registrar.cpp:461] Applied 1 operations in > 3050ns; attempting to update the registry > I1007 14:12:58.000722 28646 log.cpp:577] Attempting to append 235 bytes to > the log > I1007 14:12:58.000799 28651 coordinator.cpp:348] Coordinator attempting to > write APPEND action at position 1 > I1007 14:12:58.001106 28648 replica.cpp:537] Replica received write request > for position 1 from __req_res__(6239)@172.30.2.161:38776 > I1007 14:12:58.002013 28648 leveldb.cpp:341] Persisting action (254 bytes) to > leveldb took 884026ns > I1007 14:12:58.002027 28648 replica.cpp:708] Persisted action APPEND at > position 1 > I1007 14:12:58.002336 28650 replica.cpp:691] Replica received learned notice > for position 1 from @0.0.0.0:0 > I1007 14:12:58.004251 28650 leveldb.cpp:341] Persisting action (256 bytes) to > leveldb took 1.897221ms > I1007 14:12:58.004268 28650 replica.cpp:708] Persisted action APPEND at > position 1 > I1007 14:12:58.004483 28645 registrar.cpp:506] Successfully updated the > registry in 3.902208ms > I1007 14:12:58.004537 28646 log.cpp:596] Attempting to truncate the log to 1 > I1007 14:12:58.004566 28644 coordinator.cpp:348] Coordinator attempting to > write TRUNCATE action at position 2 > I1007 14:12:58.004580 28645 registrar.cpp:392] Successfully recovered > registrar > I1007 14:12:58.004871 28650 master.cpp:1676] Recovered 0 agents from the > registry (196B); allowing 10mins for agents to re-register > I1007 14:12:58.004896 28648 replica.cpp:537] Replica received write request > for position 2 from __req_res__(6240)@172.30.2.161:38776 > I1007 14:12:58.004914 28649 hierarchical.cpp:176] Skipping recovery of > hierarchical allocator: nothing to recover > I1007 14:12:58.006487 28648 leveldb.cpp:341] Persisting action (16 bytes) to > leveldb took 1.566923ms > I1007 14:12:58.006503 28648 replica.cpp:708] Persisted action TRUNCATE at > position 2 > I1007 14:12:58.006700 28651 replica.cpp:691] Replica received learned notice > for position 2 from @0.0.0.0:0 > I1007 14:12:58.008687 28651 leveldb.cpp:341] Persisting action (18 bytes) to > leveldb took 1.966404ms > I1007 14:12:58.008713 28651 leveldb.cpp:399] Deleting ~1 keys from leveldb > took 10225ns > I1007 14:12:58.008721 28651 replica.cpp:708] Persisted action TRUNCATE at > position 2 > I1007 14:12:58.010568 28645 slave.cpp:208] Mesos agent started on > @172.30.2.161:38776 > I1007 14:12:58.010679 28630 scheduler.cpp:176] Version: 1.1.0 > I1007 14:12:58.010578 28645 slave.cpp:209] Flags at startup: --acls="" > --appc_simple_discovery_uri_prefix="http://" > --appc_store_dir="/tmp/mesos/store/appc" --authenticate_http_readonly="true" > --authenticate_http_readwrite="true" --authenticatee="crammd5" > --authentication_backoff_factor="1secs" --authorizer="local" > --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" > --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" > --cgroups_root="mesos" --container_disk_watch_interval="15secs" > --containerizers="mesos" > --credential="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/credential" > --default_role="*" --disk_watch_interval="1mins" --docker="docker" > --docker_kill_orphans="true" --docker_registry="https://registry-1.docker.io" > --docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock" > --docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker" > --docker_volume_checkpoint_dir="/var/run/mesos/isolators/docker/volume" > --enforce_container_disk_quota="false" > --executor_registration_timeout="1mins" > --executor_shutdown_grace_period="5secs" > --fetcher_cache_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/fetch" > --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="1weeks" > --gc_disk_headroom="0.1" --hadoop_home="" --help="false" > --hostname_lookup="true" --http_authenticators="basic" > --http_command_executor="false" > --http_credentials="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/http_credentials" > --image_provisioner_backend="copy" --initialize_driver_logging="true" > --isolation="posix/cpu,posix/mem" --launcher="linux" > --launcher_dir="/mnt/teamcity/work/4240ba9ddd0997c3/build/src" > --logbufsecs="0" --logging_level="INFO" > --oversubscribed_resources_interval="15secs" --perf_duration="10secs" > --perf_interval="1mins" --qos_correction_interval_min="0ns" --quiet="false" > --recover="reconnect" --recovery_timeout="15mins" > --registration_backoff_factor="10ms" > --resources="cpus:2;gpus:0;mem:1024;disk:1024;ports:[31000-32000]" > --revocable_cpu_low_priority="true" > --runtime_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5" > --sandbox_directory="/mnt/mesos/sandbox" --strict="true" > --switch_user="true" --systemd_enable_support="true" > --systemd_runtime_directory="/run/systemd/system" --version="false" > --work_dir="/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW" > --xfs_project_range="[5000-10000]" > I1007 14:12:58.010781 28645 credentials.hpp:86] Loading credential for > authentication from > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/credential' > I1007 14:12:58.010834 28645 slave.cpp:346] Agent using credential for: > test-principal > I1007 14:12:58.010850 28645 credentials.hpp:37] Loading credentials for > authentication from > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_dv7FC5/http_credentials' > I1007 14:12:58.010880 28644 scheduler.cpp:465] New master detected at > master@172.30.2.161:38776 > I1007 14:12:58.010898 28644 scheduler.cpp:474] Waiting for 0ns before > initiating a re-(connection) attempt with the master > I1007 14:12:58.010920 28645 http.cpp:883] Using default 'basic' HTTP > authenticator for realm 'mesos-agent-readonly' > I1007 14:12:58.010948 28645 http.cpp:883] Using default 'basic' HTTP > authenticator for realm 'mesos-agent-readwrite' > I1007 14:12:58.011189 28645 slave.cpp:533] Agent resources: cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000] > I1007 14:12:58.011216 28645 slave.cpp:541] Agent attributes: [ ] > I1007 14:12:58.011221 28645 slave.cpp:546] Agent hostname: > ip-172-30-2-161.ec2.internal.mesosphere.io > I1007 14:12:58.011471 28651 state.cpp:57] Recovering state from > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/meta' > I1007 14:12:58.011592 28644 status_update_manager.cpp:203] Recovering status > update manager > I1007 14:12:58.011668 28644 slave.cpp:5256] Finished recovery > I1007 14:12:58.011844 28644 slave.cpp:5428] Querying resource estimator for > oversubscribable resources > I1007 14:12:58.011942 28644 slave.cpp:915] New master detected at > master@172.30.2.161:38776 > I1007 14:12:58.011958 28651 status_update_manager.cpp:177] Pausing sending > status updates > I1007 14:12:58.011960 28644 slave.cpp:974] Authenticating with master > master@172.30.2.161:38776 > I1007 14:12:58.011991 28644 slave.cpp:985] Using default CRAM-MD5 > authenticatee > I1007 14:12:58.012025 28644 slave.cpp:947] Detecting new master > I1007 14:12:58.012051 28649 authenticatee.cpp:121] Creating new client SASL > connection > I1007 14:12:58.012137 28651 master.cpp:6634] Authenticating > (401)@172.30.2.161:38776 > I1007 14:12:58.012226 28647 authenticator.cpp:414] Starting authentication > session for crammd5-authenticatee(991)@172.30.2.161:38776 > I1007 14:12:58.012310 28651 authenticator.cpp:98] Creating new server SASL > connection > I1007 14:12:58.012409 28651 authenticatee.cpp:213] Received SASL > authentication mechanisms: CRAM-MD5 > I1007 14:12:58.012428 28651 authenticatee.cpp:239] Attempting to authenticate > with mechanism 'CRAM-MD5' > I1007 14:12:58.012471 28651 authenticator.cpp:204] Received SASL > authentication start > I1007 14:12:58.012507 28651 authenticator.cpp:326] Authentication requires > more steps > I1007 14:12:58.012538 28648 scheduler.cpp:353] Connected with the master at > http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.012552 28647 authenticatee.cpp:259] Received SASL > authentication step > I1007 14:12:58.012673 28651 authenticator.cpp:232] Received SASL > authentication step > I1007 14:12:58.012693 28651 auxprop.cpp:109] Request to lookup properties for > user: 'test-principal' realm: 'ip-172-30-2-161.ec2.internal' server FQDN: > 'ip-172-30-2-161.ec2.internal' SASL_AUXPROP_VERIFY_AGAINST_HASH: false > SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: false > I1007 14:12:58.012701 28651 auxprop.cpp:181] Looking up auxiliary property > '*userPassword' > I1007 14:12:58.012712 28651 auxprop.cpp:181] Looking up auxiliary property > '*cmusaslsecretCRAM-MD5' > I1007 14:12:58.012722 28651 auxprop.cpp:109] Request to lookup properties for > user: 'test-principal' realm: 'ip-172-30-2-161.ec2.internal' server FQDN: > 'ip-172-30-2-161.ec2.internal' SASL_AUXPROP_VERIFY_AGAINST_HASH: false > SASL_AUXPROP_OVERRIDE: false SASL_AUXPROP_AUTHZID: true > I1007 14:12:58.012729 28651 auxprop.cpp:131] Skipping auxiliary property > '*userPassword' since SASL_AUXPROP_AUTHZID == true > I1007 14:12:58.012737 28651 auxprop.cpp:131] Skipping auxiliary property > '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true > I1007 14:12:58.012749 28651 authenticator.cpp:318] Authentication success > I1007 14:12:58.012830 28648 authenticatee.cpp:299] Authentication success > I1007 14:12:58.012850 28645 master.cpp:6664] Successfully authenticated > principal 'test-principal' at (401)@172.30.2.161:38776 > I1007 14:12:58.012876 28646 authenticator.cpp:432] Authentication session > cleanup for crammd5-authenticatee(991)@172.30.2.161:38776 > I1007 14:12:58.012972 28648 slave.cpp:1069] Successfully authenticated with > master master@172.30.2.161:38776 > I1007 14:12:58.013017 28651 scheduler.cpp:235] Sending SUBSCRIBE call to > http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.013093 28648 slave.cpp:1475] Will retry registration in > 4.168895ms if necessary > I1007 14:12:58.013185 28649 master.cpp:5074] Registering agent at > (401)@172.30.2.161:38776 (ip-172-30-2-161.ec2.internal.mesosphere.io) with id > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 > I1007 14:12:58.013326 28649 process.cpp:3377] Handling HTTP event for process > 'master' with path: '/master/api/v1/scheduler' > I1007 14:12:58.013386 28644 registrar.cpp:461] Applied 1 operations in > 14823ns; attempting to update the registry > I1007 14:12:58.013556 28651 log.cpp:577] Attempting to append 434 bytes to > the log > I1007 14:12:58.013615 28651 coordinator.cpp:348] Coordinator attempting to > write APPEND action at position 3 > I1007 14:12:58.013741 28650 http.cpp:382] HTTP POST for > /master/api/v1/scheduler from 172.30.2.161:48518 > I1007 14:12:58.013804 28650 master.cpp:2309] Received subscription request > for HTTP framework 'default' > I1007 14:12:58.013823 28650 master.cpp:2049] Authorizing framework principal > 'test-principal' to receive offers for role '*' > I1007 14:12:58.013964 28648 master.cpp:2407] Subscribing framework 'default' > with checkpointing disabled and capabilities [ ] > I1007 14:12:58.013985 28649 replica.cpp:537] Replica received write request > for position 3 from __req_res__(6241)@172.30.2.161:38776 > I1007 14:12:58.014142 28648 master.hpp:2162] Sending heartbeat to > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.014166 28644 hierarchical.cpp:275] Added framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.014189 28644 hierarchical.cpp:1694] No allocations performed > I1007 14:12:58.014200 28644 hierarchical.cpp:1789] No inverse offers to send > out! > I1007 14:12:58.014214 28644 hierarchical.cpp:1286] Performed allocation for 0 > agents in 31808ns > I1007 14:12:58.014370 28645 scheduler.cpp:671] Enqueuing event SUBSCRIBED > received from http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.014456 28649 leveldb.cpp:341] Persisting action (453 bytes) to > leveldb took 448489ns > I1007 14:12:58.014470 28649 replica.cpp:708] Persisted action APPEND at > position 3 > I1007 14:12:58.014523 28648 scheduler.cpp:671] Enqueuing event HEARTBEAT > received from http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.014883 28645 replica.cpp:691] Replica received learned notice > for position 3 from @0.0.0.0:0 > I1007 14:12:58.016865 28645 leveldb.cpp:341] Persisting action (455 bytes) to > leveldb took 1.963645ms > I1007 14:12:58.016880 28645 replica.cpp:708] Persisted action APPEND at > position 3 > I1007 14:12:58.017127 28644 registrar.cpp:506] Successfully updated the > registry in 3.705088ms > I1007 14:12:58.017184 28648 log.cpp:596] Attempting to truncate the log to 3 > I1007 14:12:58.017242 28651 coordinator.cpp:348] Coordinator attempting to > write TRUNCATE action at position 4 > I1007 14:12:58.017367 28645 master.cpp:5145] Registered agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) with cpus(*):2; mem(*):1024; > disk(*):1024; ports(*):[31000-32000] > I1007 14:12:58.017423 28644 slave.cpp:4108] Received ping from > slave-observer(454)@172.30.2.161:38776 > I1007 14:12:58.017436 28650 hierarchical.cpp:485] Added agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 > (ip-172-30-2-161.ec2.internal.mesosphere.io) with cpus(*):2; mem(*):1024; > disk(*):1024; ports(*):[31000-32000] (allocated: {}) > I1007 14:12:58.017483 28644 slave.cpp:1115] Registered with master > master@172.30.2.161:38776; given agent ID > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 > I1007 14:12:58.017501 28644 fetcher.cpp:86] Clearing fetcher cache > I1007 14:12:58.017575 28651 status_update_manager.cpp:184] Resuming sending > status updates > I1007 14:12:58.017680 28650 hierarchical.cpp:1789] No inverse offers to send > out! > I1007 14:12:58.017709 28650 hierarchical.cpp:1309] Performed allocation for > agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 in 246743ns > I1007 14:12:58.017750 28644 slave.cpp:1138] Checkpointing SlaveInfo to > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/meta/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/slave.info' > I1007 14:12:58.017774 28645 master.cpp:6463] Sending 1 offers to framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default) > I1007 14:12:58.017982 28644 replica.cpp:537] Replica received write request > for position 4 from __req_res__(6242)@172.30.2.161:38776 > I1007 14:12:58.018200 28645 scheduler.cpp:671] Enqueuing event OFFERS > received from http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.018880 28646 scheduler.cpp:235] Sending ACCEPT call to > http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.019086 28644 leveldb.cpp:341] Persisting action (16 bytes) to > leveldb took 1.082078ms > I1007 14:12:58.019104 28644 replica.cpp:708] Persisted action TRUNCATE at > position 4 > I1007 14:12:58.019212 28645 process.cpp:3377] Handling HTTP event for process > 'master' with path: '/master/api/v1/scheduler' > I1007 14:12:58.019390 28645 replica.cpp:691] Replica received learned notice > for position 4 from @0.0.0.0:0 > I1007 14:12:58.019577 28647 http.cpp:382] HTTP POST for > /master/api/v1/scheduler from 172.30.2.161:48516 > I1007 14:12:58.019711 28647 master.cpp:3521] Processing ACCEPT call for > offers: [ 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-O0 ] on agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) for framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default) > I1007 14:12:58.019737 28647 master.cpp:3143] Authorizing framework principal > 'test-principal' to launch task ed09b29b-d149-46c1-a31d-f8fe3fdf214e > I1007 14:12:58.019785 28647 master.cpp:3143] Authorizing framework principal > 'test-principal' to launch task 9ce218ba-84b7-4ea4-a560-b30472f4a22c > I1007 14:12:58.020279 28647 master.cpp:8214] Adding task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e with resources cpus(*):0.1; mem(*):32; > disk(*):32 on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 > (ip-172-30-2-161.ec2.internal.mesosphere.io) > I1007 14:12:58.020342 28647 master.cpp:8214] Adding task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c with resources cpus(*):0.1; mem(*):32; > disk(*):32 on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 > (ip-172-30-2-161.ec2.internal.mesosphere.io) > I1007 14:12:58.020375 28647 master.cpp:4358] Launching task group { > 9ce218ba-84b7-4ea4-a560-b30472f4a22c, ed09b29b-d149-46c1-a31d-f8fe3fdf214e } > of framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default) with > resources cpus(*):0.2; mem(*):64; disk(*):64 on agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) > I1007 14:12:58.020480 28648 slave.cpp:1539] Got assigned task group > containing tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, > 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] for framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.020515 28650 hierarchical.cpp:1018] Recovered cpus(*):1.8; > mem(*):960; disk(*):960; ports(*):[31000-32000] (total: cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000], allocated: cpus(*):0.2; > mem(*):64; disk(*):64) on agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from > framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.020541 28650 hierarchical.cpp:1055] Framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 filtered agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 for 5secs > I1007 14:12:58.020830 28644 scheduler.cpp:235] Sending KILL call to > http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.021075 28651 process.cpp:3377] Handling HTTP event for process > 'master' with path: '/master/api/v1/scheduler' > I1007 14:12:58.021347 28648 http.cpp:382] HTTP POST for > /master/api/v1/scheduler from 172.30.2.161:48516 > I1007 14:12:58.021399 28648 master.cpp:4682] Telling agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) to kill task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (default) > I1007 14:12:58.021473 28646 slave.cpp:2283] Asked to kill task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > W1007 14:12:58.021491 28646 slave.cpp:2324] Killing task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 before it was launched > I1007 14:12:58.021483 28645 leveldb.cpp:341] Persisting action (18 bytes) to > leveldb took 2.0668ms > I1007 14:12:58.021524 28645 leveldb.cpp:399] Deleting ~2 keys from leveldb > took 17427ns > I1007 14:12:58.021539 28645 replica.cpp:708] Persisted action TRUNCATE at > position 4 > I1007 14:12:58.021720 28630 slave.cpp:1696] Launching task group containing > tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, > 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] for framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > W1007 14:12:58.021749 28630 slave.cpp:1736] Ignoring running task group > containing tasks [ ed09b29b-d149-46c1-a31d-f8fe3fdf214e, > 9ce218ba-84b7-4ea4-a560-b30472f4a22c ] of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 because it has been killed in the > meantime > I1007 14:12:58.021777 28630 slave.cpp:3609] Handling status update > TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from @0.0.0.0:0 > W1007 14:12:58.021800 28630 slave.cpp:3705] Could not find the executor for > status update TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for > task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.021888 28630 slave.cpp:3609] Handling status update > TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from @0.0.0.0:0 > I1007 14:12:58.021890 28647 status_update_manager.cpp:323] Received status > update TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.021919 28647 status_update_manager.cpp:500] Creating > StatusUpdate stream for task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of > framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > W1007 14:12:58.021913 28630 slave.cpp:3705] Could not find the executor for > status update TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for > task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.021981 28630 slave.cpp:4709] Cleaning up framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.021986 28647 status_update_manager.cpp:377] Forwarding update > TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to the agent > E1007 14:12:58.022033 28630 slave.cpp:5408] Failed to find the mtime of > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/frameworks/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000': > Error invoking stat for > '/mnt/teamcity/temp/buildTmp/SlaveTest_KillTaskGroupBetweenRunTaskParts_UtMqTW/slaves/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0/frameworks/76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000': > No such file or directory > I1007 14:12:58.022058 28651 slave.cpp:4026] Forwarding the update TASK_KILLED > (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to master@172.30.2.161:38776 > I1007 14:12:58.022063 28647 status_update_manager.cpp:323] Received status > update TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022099 28647 status_update_manager.cpp:500] Creating > StatusUpdate stream for task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of > framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022140 28651 slave.cpp:3920] Status update manager > successfully handled status update TASK_KILLED (UUID: > b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022178 28647 status_update_manager.cpp:377] Forwarding update > TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to the agent > I1007 14:12:58.022231 28646 master.cpp:5672] Status update TASK_KILLED (UUID: > b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) > I1007 14:12:58.022239 28651 slave.cpp:4026] Forwarding the update TASK_KILLED > (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 to master@172.30.2.161:38776 > I1007 14:12:58.022256 28646 master.cpp:5734] Forwarding status update > TASK_KILLED (UUID: b28bbec8-5746-4f3a-bf53-2e4f1a78b4cb) for task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022279 28647 status_update_manager.cpp:285] Closing status > update streams for framework 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022294 28647 status_update_manager.cpp:531] Cleaning up status > update stream for task 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022301 28651 slave.cpp:3920] Status update manager > successfully handled status update TASK_KILLED (UUID: > a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022330 28647 status_update_manager.cpp:531] Cleaning up status > update stream for task ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022347 28646 master.cpp:7592] Updating the state of task > ed09b29b-d149-46c1-a31d-f8fe3fdf214e of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (latest state: TASK_KILLED, status > update state: TASK_KILLED) > I1007 14:12:58.022449 28646 master.cpp:5672] Status update TASK_KILLED (UUID: > a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 from agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 at (401)@172.30.2.161:38776 > (ip-172-30-2-161.ec2.internal.mesosphere.io) > I1007 14:12:58.022464 28646 master.cpp:5734] Forwarding status update > TASK_KILLED (UUID: a768cdc6-4d09-467c-8a66-f0117744a8dd) for task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022461 28644 hierarchical.cpp:1018] Recovered cpus(*):0.1; > mem(*):32; disk(*):32 (total: cpus(*):2; mem(*):1024; disk(*):1024; > ports(*):[31000-32000], allocated: cpus(*):0.1; mem(*):32; disk(*):32) on > agent 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022600 28650 scheduler.cpp:671] Enqueuing event UPDATE > received from http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.022609 28646 master.cpp:7592] Updating the state of task > 9ce218ba-84b7-4ea4-a560-b30472f4a22c of framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 (latest state: TASK_KILLED, status > update state: TASK_KILLED) > I1007 14:12:58.022825 28644 hierarchical.cpp:1018] Recovered cpus(*):0.1; > mem(*):32; disk(*):32 (total: cpus(*):2; mem(*):1024; disk(*):1024; > ports(*):[31000-32000], allocated: {}) on agent > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-S0 from framework > 76d4d55f-dcc6-4033-85d9-7ec97ef353cb-0000 > I1007 14:12:58.022981 28645 scheduler.cpp:671] Enqueuing event UPDATE > received from http://172.30.2.161:38776/master/api/v1/scheduler > I1007 14:12:58.023365 28651 slave.cpp:787] Agent terminating > *** Aborted at 1475849578 (unix time) try "date -d @1475849578" if you are > using GNU date *** > PC: @ 0x7fdbf26dc95f mesos::internal::slave::Slave::finalize() > *** SIGSEGV (@0x88) received by PID 28630 (TID 0x7fdbe5910700) from PID 136; > stack trace: *** > @ 0x7fdbdef6fbb2 (unknown) > @ 0x7fdbdef74009 (unknown) > @ 0x7fdbdef67f08 (unknown) > @ 0x7fdbf14bc9f0 (unknown) > @ 0x7fdbf26dc95f mesos::internal::slave::Slave::finalize() > @ 0x7fdbf2ec461b process::ProcessManager::resume() > @ 0x7fdbf2ec7cb7 > _ZNSt6thread5_ImplISt12_Bind_simpleIFZN7process14ProcessManager12init_threadsEvEUt_vEEE6_M_runEv > @ 0x7fdbf0dd6f20 (unknown) > @ 0x7fdbf14b360a start_thread > @ 0x7fdbf054659d __clone > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)