[ https://issues.apache.org/jira/browse/MESOS-1653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Benjamin Mahler resolved MESOS-1653. ------------------------------------ Resolution: Fixed Fix Version/s: 0.20.0 Assignee: Timothy Chen Fix was included here: {noformat} commit 656b0e075c79e03cf6937bbe7302424768729aa2 Author: Timothy Chen <tnac...@apache.org> Date: Wed Aug 6 11:34:03 2014 -0700 Re-enabled HealthCheckTest.ConsecutiveFailures test. The test originally was flaky because the time to process the number of consecutive checks configured exceeds the task itself, so the task finished but the number of expected task health check didn't match. Review: https://reviews.apache.org/r/23772 {noformat} > HealthCheckTest.GracePeriod is flaky. > ------------------------------------- > > Key: MESOS-1653 > URL: https://issues.apache.org/jira/browse/MESOS-1653 > Project: Mesos > Issue Type: Bug > Components: test > Reporter: Benjamin Mahler > Assignee: Timothy Chen > Fix For: 0.20.0 > > > {noformat} > [----------] 3 tests from HealthCheckTest > [ RUN ] HealthCheckTest.GracePeriod > Using temporary directory '/tmp/HealthCheckTest_GracePeriod_d7zCPr' > I0729 17:10:10.484951 1176 leveldb.cpp:176] Opened db in 28.883552ms > I0729 17:10:10.499487 1176 leveldb.cpp:183] Compacted db in 13.674118ms > I0729 17:10:10.500200 1176 leveldb.cpp:198] Created db iterator in 7394ns > I0729 17:10:10.500692 1176 leveldb.cpp:204] Seeked to beginning of db in > 2317ns > I0729 17:10:10.501113 1176 leveldb.cpp:273] Iterated through 0 keys in the > db in 1367ns > I0729 17:10:10.501535 1176 replica.cpp:741] Replica recovered with log > positions 0 -> 0 with 1 holes and 0 unlearned > I0729 17:10:10.502233 1212 recover.cpp:425] Starting replica recovery > I0729 17:10:10.502295 1212 recover.cpp:451] Replica is in EMPTY status > I0729 17:10:10.502825 1212 replica.cpp:638] Replica in EMPTY status received > a broadcasted recover request > I0729 17:10:10.502877 1212 recover.cpp:188] Received a recover response from > a replica in EMPTY status > I0729 17:10:10.502980 1212 recover.cpp:542] Updating replica status to > STARTING > I0729 17:10:10.508482 1213 master.cpp:289] Master > 20140729-171010-16842879-54701-1176 (trusty) started on 127.0.1.1:54701 > I0729 17:10:10.508607 1213 master.cpp:326] Master only allowing > authenticated frameworks to register > I0729 17:10:10.508632 1213 master.cpp:331] Master only allowing > authenticated slaves to register > I0729 17:10:10.508656 1213 credentials.hpp:36] Loading credentials for > authentication from '/tmp/HealthCheckTest_GracePeriod_d7zCPr/credentials' > I0729 17:10:10.509407 1213 master.cpp:360] Authorization enabled > I0729 17:10:10.510030 1207 hierarchical_allocator_process.hpp:301] > Initializing hierarchical allocator process with master : > master@127.0.1.1:54701 > I0729 17:10:10.510113 1207 master.cpp:123] No whitelist given. Advertising > offers for all slaves > I0729 17:10:10.511699 1213 master.cpp:1129] The newly elected leader is > master@127.0.1.1:54701 with id 20140729-171010-16842879-54701-1176 > I0729 17:10:10.512230 1213 master.cpp:1142] Elected as the leading master! > I0729 17:10:10.512692 1213 master.cpp:960] Recovering from registrar > I0729 17:10:10.513226 1210 registrar.cpp:313] Recovering registrar > I0729 17:10:10.516006 1212 leveldb.cpp:306] Persisting metadata (8 bytes) to > leveldb took 12.946461ms > I0729 17:10:10.516047 1212 replica.cpp:320] Persisted replica status to > STARTING > I0729 17:10:10.516129 1212 recover.cpp:451] Replica is in STARTING status > I0729 17:10:10.516520 1212 replica.cpp:638] Replica in STARTING status > received a broadcasted recover request > I0729 17:10:10.516592 1212 recover.cpp:188] Received a recover response from > a replica in STARTING status > I0729 17:10:10.516767 1212 recover.cpp:542] Updating replica status to VOTING > I0729 17:10:10.528376 1212 leveldb.cpp:306] Persisting metadata (8 bytes) to > leveldb took 11.537102ms > I0729 17:10:10.528430 1212 replica.cpp:320] Persisted replica status to > VOTING > I0729 17:10:10.528501 1212 recover.cpp:556] Successfully joined the Paxos > group > I0729 17:10:10.528565 1212 recover.cpp:440] Recover process terminated > I0729 17:10:10.528700 1212 log.cpp:656] Attempting to start the writer > I0729 17:10:10.528960 1212 replica.cpp:474] Replica received implicit > promise request with proposal 1 > I0729 17:10:10.537821 1212 leveldb.cpp:306] Persisting metadata (8 bytes) to > leveldb took 8.830863ms > I0729 17:10:10.537869 1212 replica.cpp:342] Persisted promised to 1 > I0729 17:10:10.540550 1209 coordinator.cpp:230] Coordinator attemping to > fill missing position > I0729 17:10:10.540856 1209 replica.cpp:375] Replica received explicit > promise request for position 0 with proposal 2 > I0729 17:10:10.547430 1209 leveldb.cpp:343] Persisting action (8 bytes) to > leveldb took 6.548344ms > I0729 17:10:10.547471 1209 replica.cpp:676] Persisted action at 0 > I0729 17:10:10.547732 1209 replica.cpp:508] Replica received write request > for position 0 > I0729 17:10:10.547765 1209 leveldb.cpp:438] Reading position from leveldb > took 15676ns > I0729 17:10:10.557169 1209 leveldb.cpp:343] Persisting action (14 bytes) to > leveldb took 9.373798ms > I0729 17:10:10.557241 1209 replica.cpp:676] Persisted action at 0 > I0729 17:10:10.560642 1210 replica.cpp:655] Replica received learned notice > for position 0 > I0729 17:10:10.570312 1210 leveldb.cpp:343] Persisting action (16 bytes) to > leveldb took 9.61503ms > I0729 17:10:10.570380 1210 replica.cpp:676] Persisted action at 0 > I0729 17:10:10.570406 1210 replica.cpp:661] Replica learned NOP action at > position 0 > I0729 17:10:10.570746 1210 log.cpp:672] Writer started with ending position 0 > I0729 17:10:10.571141 1210 leveldb.cpp:438] Reading position from leveldb > took 24085ns > I0729 17:10:10.580646 1210 registrar.cpp:346] Successfully fetched the > registry (0B) > I0729 17:10:10.580708 1210 registrar.cpp:422] Attempting to update the > 'registry' > I0729 17:10:10.582347 1209 log.cpp:680] Attempting to append 118 bytes to > the log > I0729 17:10:10.582406 1209 coordinator.cpp:340] Coordinator attempting to > write APPEND action at position 1 > I0729 17:10:10.582655 1209 replica.cpp:508] Replica received write request > for position 1 > I0729 17:10:10.591322 1209 leveldb.cpp:343] Persisting action (135 bytes) to > leveldb took 8.628332ms > I0729 17:10:10.591384 1209 replica.cpp:676] Persisted action at 1 > I0729 17:10:10.591640 1209 replica.cpp:655] Replica received learned notice > for position 1 > I0729 17:10:10.601271 1209 leveldb.cpp:343] Persisting action (137 bytes) to > leveldb took 9.600283ms > I0729 17:10:10.601327 1209 replica.cpp:676] Persisted action at 1 > I0729 17:10:10.601341 1209 replica.cpp:661] Replica learned APPEND action at > position 1 > I0729 17:10:10.608666 1209 registrar.cpp:479] Successfully updated 'registry' > I0729 17:10:10.608739 1209 registrar.cpp:372] Successfully recovered > registrar > I0729 17:10:10.608855 1209 master.cpp:987] Recovered 0 slaves from the > Registry (82B) ; allowing 10mins for slaves to re-register > I0729 17:10:10.608901 1209 log.cpp:699] Attempting to truncate the log to 1 > I0729 17:10:10.608949 1209 coordinator.cpp:340] Coordinator attempting to > write TRUNCATE action at position 2 > I0729 17:10:10.609171 1209 replica.cpp:508] Replica received write request > for position 2 > I0729 17:10:10.618360 1209 leveldb.cpp:343] Persisting action (16 bytes) to > leveldb took 9.157424ms > I0729 17:10:10.618412 1209 replica.cpp:676] Persisted action at 2 > I0729 17:10:10.618633 1209 replica.cpp:655] Replica received learned notice > for position 2 > I0729 17:10:10.627727 1209 leveldb.cpp:343] Persisting action (18 bytes) to > leveldb took 9.062752ms > I0729 17:10:10.628556 1209 leveldb.cpp:401] Deleting ~1 keys from leveldb > took 58420ns > I0729 17:10:10.628873 1209 replica.cpp:676] Persisted action at 2 > I0729 17:10:10.628903 1209 replica.cpp:661] Replica learned TRUNCATE action > at position 2 > I0729 17:10:10.640494 1176 containerizer.cpp:124] Using isolation: > posix/cpu,posix/mem > I0729 17:10:10.644561 1205 slave.cpp:169] Slave started on > 29)@127.0.1.1:54701 > I0729 17:10:10.644593 1205 credentials.hpp:84] Loading credential for > authentication from '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/credential' > I0729 17:10:10.644702 1205 slave.cpp:267] Slave using credential for: > test-principal > I0729 17:10:10.644851 1205 slave.cpp:280] Slave resources: cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000] > I0729 17:10:10.644940 1205 slave.cpp:325] Slave hostname: trusty > I0729 17:10:10.644954 1205 slave.cpp:326] Slave checkpoint: false > I0729 17:10:10.645478 1205 state.cpp:33] Recovering state from > '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/meta' > I0729 17:10:10.645634 1205 status_update_manager.cpp:193] Recovering status > update manager > I0729 17:10:10.645722 1205 containerizer.cpp:287] Recovering containerizer > I0729 17:10:10.646106 1205 slave.cpp:3128] Finished recovery > I0729 17:10:10.646536 1205 slave.cpp:601] New master detected at > master@127.0.1.1:54701 > I0729 17:10:10.646589 1205 slave.cpp:677] Authenticating with master > master@127.0.1.1:54701 > I0729 17:10:10.646651 1205 slave.cpp:650] Detecting new master > I0729 17:10:10.646694 1205 status_update_manager.cpp:167] New master > detected at master@127.0.1.1:54701 > I0729 17:10:10.646770 1205 authenticatee.hpp:128] Creating new client SASL > connection > I0729 17:10:10.647045 1205 master.cpp:3504] Authenticating > slave(29)@127.0.1.1:54701 > I0729 17:10:10.647186 1205 authenticator.hpp:156] Creating new server SASL > connection > I0729 17:10:10.647390 1205 authenticatee.hpp:219] Received SASL > authentication mechanisms: CRAM-MD5 > I0729 17:10:10.647426 1205 authenticatee.hpp:245] Attempting to authenticate > with mechanism 'CRAM-MD5' > I0729 17:10:10.647460 1205 authenticator.hpp:262] Received SASL > authentication start > I0729 17:10:10.647511 1205 authenticator.hpp:384] Authentication requires > more steps > I0729 17:10:10.647543 1205 authenticatee.hpp:265] Received SASL > authentication step > I0729 17:10:10.647629 1205 authenticator.hpp:290] Received SASL > authentication step > I0729 17:10:10.647678 1205 auxprop.cpp:81] Request to lookup properties for > user: 'test-principal' realm: 'trusty' server FQDN: 'trusty' > SASL_AUXPROP_VERIFY_AGAINST_HASH: false SASL_AUXPROP_OVERRIDE: false > SASL_AUXPROP_AUTHZID: false > I0729 17:10:10.647706 1205 auxprop.cpp:153] Looking up auxiliary property > '*userPassword' > I0729 17:10:10.647721 1205 auxprop.cpp:153] Looking up auxiliary property > '*cmusaslsecretCRAM-MD5' > I0729 17:10:10.647765 1205 auxprop.cpp:81] Request to lookup properties for > user: 'test-principal' realm: 'trusty' server FQDN: 'trusty' > SASL_AUXPROP_VERIFY_AGAINST_HASH: false SASL_AUXPROP_OVERRIDE: false > SASL_AUXPROP_AUTHZID: true > I0729 17:10:10.647776 1205 auxprop.cpp:103] Skipping auxiliary property > '*userPassword' since SASL_AUXPROP_AUTHZID == true > I0729 17:10:10.647783 1205 auxprop.cpp:103] Skipping auxiliary property > '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true > I0729 17:10:10.647806 1205 authenticator.hpp:376] Authentication success > I0729 17:10:10.647850 1205 authenticatee.hpp:305] Authentication success > I0729 17:10:10.647874 1205 master.cpp:3544] Successfully authenticated > principal 'test-principal' at slave(29)@127.0.1.1:54701 > I0729 17:10:10.647960 1205 slave.cpp:734] Successfully authenticated with > master master@127.0.1.1:54701 > I0729 17:10:10.648056 1205 slave.cpp:972] Will retry registration in > 11.218662ms if necessary > I0729 17:10:10.648203 1210 master.cpp:2762] Registering slave at > slave(29)@127.0.1.1:54701 (trusty) with id > 20140729-171010-16842879-54701-1176-0 > I0729 17:10:10.648386 1204 registrar.cpp:422] Attempting to update the > 'registry' > I0729 17:10:10.649412 1212 log.cpp:680] Attempting to append 290 bytes to > the log > I0729 17:10:10.649459 1212 coordinator.cpp:340] Coordinator attempting to > write APPEND action at position 3 > I0729 17:10:10.649659 1212 replica.cpp:508] Replica received write request > for position 3 > I0729 17:10:10.652601 1176 sched.cpp:139] Version: 0.20.0 > I0729 17:10:10.653542 1209 sched.cpp:235] New master detected at > master@127.0.1.1:54701 > I0729 17:10:10.653595 1209 sched.cpp:285] Authenticating with master > master@127.0.1.1:54701 > I0729 17:10:10.653727 1209 authenticatee.hpp:128] Creating new client SASL > connection > I0729 17:10:10.653970 1209 master.cpp:3504] Authenticating > scheduler-f3043154-6292-421a-800f-6264c426503c@127.0.1.1:54701 > I0729 17:10:10.654096 1209 authenticator.hpp:156] Creating new server SASL > connection > I0729 17:10:10.654660 1209 authenticatee.hpp:219] Received SASL > authentication mechanisms: CRAM-MD5 > I0729 17:10:10.654700 1209 authenticatee.hpp:245] Attempting to authenticate > with mechanism 'CRAM-MD5' > I0729 17:10:10.654752 1209 authenticator.hpp:262] Received SASL > authentication start > I0729 17:10:10.655232 1209 authenticator.hpp:384] Authentication requires > more steps > I0729 17:10:10.655627 1213 authenticatee.hpp:265] Received SASL > authentication step > I0729 17:10:10.656039 1209 authenticator.hpp:290] Received SASL > authentication step > I0729 17:10:10.656082 1209 auxprop.cpp:81] Request to lookup properties for > user: 'test-principal' realm: 'trusty' server FQDN: 'trusty' > SASL_AUXPROP_VERIFY_AGAINST_HASH: false SASL_AUXPROP_OVERRIDE: false > SASL_AUXPROP_AUTHZID: false > I0729 17:10:10.656105 1209 auxprop.cpp:153] Looking up auxiliary property > '*userPassword' > I0729 17:10:10.656131 1209 auxprop.cpp:153] Looking up auxiliary property > '*cmusaslsecretCRAM-MD5' > I0729 17:10:10.656157 1209 auxprop.cpp:81] Request to lookup properties for > user: 'test-principal' realm: 'trusty' server FQDN: 'trusty' > SASL_AUXPROP_VERIFY_AGAINST_HASH: false SASL_AUXPROP_OVERRIDE: false > SASL_AUXPROP_AUTHZID: true > I0729 17:10:10.656174 1209 auxprop.cpp:103] Skipping auxiliary property > '*userPassword' since SASL_AUXPROP_AUTHZID == true > I0729 17:10:10.656189 1209 auxprop.cpp:103] Skipping auxiliary property > '*cmusaslsecretCRAM-MD5' since SASL_AUXPROP_AUTHZID == true > I0729 17:10:10.656215 1209 authenticator.hpp:376] Authentication success > I0729 17:10:10.656738 1204 authenticatee.hpp:305] Authentication success > I0729 17:10:10.659857 1212 leveldb.cpp:343] Persisting action (309 bytes) to > leveldb took 10.166506ms > I0729 17:10:10.659914 1212 replica.cpp:676] Persisted action at 3 > I0729 17:10:10.660799 1207 slave.cpp:972] Will retry registration in > 8.165348ms if necessary > I0729 17:10:10.660914 1207 master.cpp:2750] Ignoring register slave message > from slave(29)@127.0.1.1:54701 (trusty) as admission is already in progress > I0729 17:10:10.661286 1212 replica.cpp:655] Replica received learned notice > for position 3 > I0729 17:10:10.662286 1213 master.cpp:3544] Successfully authenticated > principal 'test-principal' at > scheduler-f3043154-6292-421a-800f-6264c426503c@127.0.1.1:54701 > I0729 17:10:10.662789 1204 sched.cpp:359] Successfully authenticated with > master master@127.0.1.1:54701 > I0729 17:10:10.663375 1204 sched.cpp:478] Sending registration request to > master@127.0.1.1:54701 > I0729 17:10:10.664089 1207 master.cpp:1248] Received registration request > from scheduler-f3043154-6292-421a-800f-6264c426503c@127.0.1.1:54701 > I0729 17:10:10.664163 1207 master.cpp:1208] Authorizing framework principal > 'test-principal' to receive offers for role '*' > I0729 17:10:10.664638 1207 master.cpp:1307] Registering framework > 20140729-171010-16842879-54701-1176-0000 at > scheduler-f3043154-6292-421a-800f-6264c426503c@127.0.1.1:54701 > I0729 17:10:10.664856 1207 hierarchical_allocator_process.hpp:331] Added > framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.664894 1207 hierarchical_allocator_process.hpp:724] No > resources available to allocate! > I0729 17:10:10.664917 1207 hierarchical_allocator_process.hpp:686] Performed > allocation for 0 slaves in 25838ns > I0729 17:10:10.665393 1204 sched.cpp:409] Framework registered with > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.666017 1204 sched.cpp:423] Scheduler::registered took 32926ns > I0729 17:10:10.669276 1210 slave.cpp:972] Will retry registration in > 28.005974ms if necessary > I0729 17:10:10.669387 1210 master.cpp:2750] Ignoring register slave message > from slave(29)@127.0.1.1:54701 (trusty) as admission is already in progress > I0729 17:10:10.669477 1212 leveldb.cpp:343] Persisting action (311 bytes) to > leveldb took 8.158554ms > I0729 17:10:10.669513 1212 replica.cpp:676] Persisted action at 3 > I0729 17:10:10.669533 1212 replica.cpp:661] Replica learned APPEND action at > position 3 > I0729 17:10:10.670101 1212 registrar.cpp:479] Successfully updated 'registry' > I0729 17:10:10.670265 1212 log.cpp:699] Attempting to truncate the log to 3 > I0729 17:10:10.670399 1212 master.cpp:2802] Registered slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > I0729 17:10:10.670454 1212 master.cpp:3972] Adding slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > with cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000] > I0729 17:10:10.670725 1212 coordinator.cpp:340] Coordinator attempting to > write TRUNCATE action at position 4 > I0729 17:10:10.670883 1212 slave.cpp:768] Registered with master > master@127.0.1.1:54701; given slave ID 20140729-171010-16842879-54701-1176-0 > I0729 17:10:10.671022 1212 hierarchical_allocator_process.hpp:444] Added > slave 20140729-171010-16842879-54701-1176-0 (trusty) with cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000] (and cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000] available) > I0729 17:10:10.671140 1212 hierarchical_allocator_process.hpp:750] Offering > cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 to framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.671286 1212 hierarchical_allocator_process.hpp:706] Performed > allocation for slave 20140729-171010-16842879-54701-1176-0 in 201281ns > I0729 17:10:10.671383 1212 slave.cpp:2325] Received ping from > slave-observer(27)@127.0.1.1:54701 > I0729 17:10:10.671535 1212 master.hpp:816] Adding offer > 20140729-171010-16842879-54701-1176-0 with resources cpus(*):2; mem(*):1024; > disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 (trusty) > I0729 17:10:10.671619 1212 master.cpp:3451] Sending 1 offers to framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.671906 1212 sched.cpp:546] Scheduler::resourceOffers took > 48649ns > I0729 17:10:10.672193 1212 replica.cpp:508] Replica received write request > for position 4 > I0729 17:10:10.675730 1207 master.hpp:826] Removing offer > 20140729-171010-16842879-54701-1176-0 with resources cpus(*):2; mem(*):1024; > disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 (trusty) > I0729 17:10:10.675851 1207 master.cpp:2125] Processing reply for offers: [ > 20140729-171010-16842879-54701-1176-0 ] on slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > for framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.675937 1207 master.cpp:2211] Authorizing framework principal > 'test-principal' to launch task 1 as user 'jenkins' > I0729 17:10:10.676506 1207 master.hpp:788] Adding task 1 with resources > cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 (trusty) > I0729 17:10:10.676573 1207 master.cpp:2277] Launching task 1 of framework > 20140729-171010-16842879-54701-1176-0000 with resources cpus(*):2; > mem(*):1024; disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > I0729 17:10:10.676806 1207 slave.cpp:1003] Got assigned task 1 for framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.677433 1207 slave.cpp:1113] Launching task 1 for framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.679307 1205 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 34137ns > I0729 17:10:10.679386 1205 master.cpp:123] No whitelist given. Advertising > offers for all slaves > I0729 17:10:10.680003 1205 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 23267ns > I0729 17:10:10.682334 1212 leveldb.cpp:343] Persisting action (16 bytes) to > leveldb took 10.104651ms > I0729 17:10:10.682368 1212 replica.cpp:676] Persisted action at 4 > I0729 17:10:10.682564 1212 replica.cpp:655] Replica received learned notice > for position 4 > I0729 17:10:10.684147 1210 containerizer.cpp:427] Starting container > 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' for executor '1' of framework > '20140729-171010-16842879-54701-1176-0000' > I0729 17:10:10.700867 1207 slave.cpp:1223] Queuing task '1' for executor 1 > of framework '20140729-171010-16842879-54701-1176-0000 > I0729 17:10:10.702924 1212 leveldb.cpp:343] Persisting action (18 bytes) to > leveldb took 20.082347ms > I0729 17:10:10.703104 1212 leveldb.cpp:401] Deleting ~2 keys from leveldb > took 110401ns > I0729 17:10:10.703147 1212 replica.cpp:676] Persisted action at 4 > I0729 17:10:10.703191 1212 replica.cpp:661] Replica learned TRUNCATE action > at position 4 > I0729 17:10:10.704622 1207 slave.cpp:562] Successfully attached file > '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/slaves/20140729-171010-16842879-54701-1176-0/frameworks/20140729-171010-16842879-54701-1176-0000/executors/1/runs/a91c7b26-f710-4ef4-b0d8-a21c4e12b982' > I0729 17:10:10.706490 1210 launcher.cpp:137] Forked child with pid '1418' > for container 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' > I0729 17:10:10.714651 1210 containerizer.cpp:537] Fetching URIs for > container 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' using command > '/var/jenkins/workspace/mesos-ubuntu-14.04-gcc/src/mesos-fetcher' > I0729 17:10:11.679939 1209 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 74216ns > I0729 17:10:11.688122 1206 slave.cpp:2470] Monitoring executor '1' of > framework '20140729-171010-16842879-54701-1176-0000' in container > 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' > I0729 17:10:13.250228 1209 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 58601ns > I0729 17:10:13.676744 1210 slave.cpp:1734] Got registration for executor '1' > of framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.678890 1214 process.cpp:1098] Socket closed while receiving > I0729 17:10:13.679328 1210 slave.cpp:1853] Flushing queued task 1 for > executor '1' of framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.682858 1214 process.cpp:1098] Socket closed while receiving > I0729 17:10:13.689992 1205 slave.cpp:2088] Handling status update > TASK_RUNNING (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 of > framework 20140729-171010-16842879-54701-1176-0000 from > executor(1)@127.0.1.1:57713 > I0729 17:10:13.690161 1205 status_update_manager.cpp:320] Received status > update TASK_RUNNING (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 > of framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.690181 1205 status_update_manager.cpp:499] Creating > StatusUpdate stream for task 1 of framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.690289 1205 status_update_manager.cpp:373] Forwarding status > update TASK_RUNNING (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 > of framework 20140729-171010-16842879-54701-1176-0000 to > master@127.0.1.1:54701 > I0729 17:10:13.690444 1205 master.cpp:3131] Forwarding status update > TASK_RUNNING (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 of > framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.690490 1205 master.cpp:3097] Status update TASK_RUNNING > (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 of framework > 20140729-171010-16842879-54701-1176-0000 from slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > I0729 17:10:13.690525 1205 slave.cpp:2246] Status update manager > successfully handled status update TASK_RUNNING (UUID: > c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 of framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.690537 1205 slave.cpp:2252] Sending acknowledgement for > status update TASK_RUNNING (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for > task 1 of framework 20140729-171010-16842879-54701-1176-0000 to > executor(1)@127.0.1.1:57713 > I0729 17:10:13.690765 1205 sched.cpp:637] Scheduler::statusUpdate took > 51765ns > I0729 17:10:13.690829 1205 master.cpp:2612] Forwarding status update > acknowledgement c0053bfe-8014-4d12-ac27-83fb49300c05 for task 1 of framework > 20140729-171010-16842879-54701-1176-0000 to slave > 20140729-171010-16842879-54701-1176-0 at slave(29)@127.0.1.1:54701 (trusty) > I0729 17:10:13.690901 1205 status_update_manager.cpp:398] Received status > update acknowledgement (UUID: c0053bfe-8014-4d12-ac27-83fb49300c05) for task > 1 of framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.690966 1205 slave.cpp:1674] Status update manager > successfully handled status update acknowledgement (UUID: > c0053bfe-8014-4d12-ac27-83fb49300c05) for task 1 of framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:13.691596 1214 process.cpp:1098] Socket closed while receiving > I0729 17:10:15.462894 1205 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 41972ns > I0729 17:10:15.463147 1205 master.cpp:123] No whitelist given. Advertising > offers for all slaves > I0729 17:10:16.464249 1210 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 52485ns > I0729 17:10:17.465081 1205 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 63487ns > I0729 17:10:18.465644 1206 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 67852ns > I0729 17:10:19.466233 1205 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 56573ns > I0729 17:10:19.676609 1206 slave.cpp:2325] Received ping from > slave-observer(27)@127.0.1.1:54701 > I0729 17:10:20.463575 1206 master.cpp:123] No whitelist given. Advertising > offers for all slaves > I0729 17:10:20.467237 1206 hierarchical_allocator_process.hpp:686] Performed > allocation for 1 slaves in 64896ns > tests/health_check_tests.cpp:384: Failure > Failed to wait 10secs for statusHealth > tests/health_check_tests.cpp:366: Failure > Actual function call count doesn't match EXPECT_CALL(sched, > statusUpdate(&driver, _))... > Expected: to be called at least twice > Actual: called once - unsatisfied and active > I0729 17:10:20.818284 1207 master.cpp:717] Framework > 20140729-171010-16842879-54701-1176-0000 disconnected > I0729 17:10:20.818326 1207 master.cpp:1584] Deactivating framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.818362 1207 master.cpp:739] Giving framework > 20140729-171010-16842879-54701-1176-0000 0ns to failover > I0729 17:10:20.818475 1207 hierarchical_allocator_process.hpp:407] > Deactivated framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.818897 1207 master.cpp:3367] Framework failover timeout, > removing framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.818914 1207 master.cpp:3826] Removing framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.819068 1207 master.hpp:806] Removing task 1 with resources > cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000] on slave > 20140729-171010-16842879-54701-1176-0 (trusty) > W0729 17:10:20.819154 1207 master.cpp:4261] Removing task 1 of framework > 20140729-171010-16842879-54701-1176-0000 and slave > 20140729-171010-16842879-54701-1176-0 in non-terminal state TASK_RUNNING > I0729 17:10:20.819250 1207 slave.cpp:1407] Asked to shut down framework > 20140729-171010-16842879-54701-1176-0000 by master@127.0.1.1:54701 > I0729 17:10:20.819282 1207 slave.cpp:1432] Shutting down framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.819300 1207 slave.cpp:2808] Shutting down executor '1' of > framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.819830 1204 hierarchical_allocator_process.hpp:635] Recovered > cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000] (total > allocatable: cpus(*):2; mem(*):1024; disk(*):1024; ports(*):[31000-32000]) on > slave 20140729-171010-16842879-54701-1176-0 from framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.819888 1204 hierarchical_allocator_process.hpp:362] Removed > framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:20.824790 1214 process.cpp:1098] Socket closed while receiving > I0729 17:10:20.826246 1209 master.cpp:626] Master terminating > I0729 17:10:20.833003 1205 slave.cpp:2332] master@127.0.1.1:54701 exited > W0729 17:10:20.833041 1205 slave.cpp:2335] Master disconnected! Waiting for > a new master to be elected > I0729 17:10:20.840898 1207 containerizer.cpp:903] Destroying container > 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' > I0729 17:10:20.973546 1214 process.cpp:1037] Socket closed while receiving > I0729 17:10:21.469235 1213 containerizer.cpp:1019] Executor for container > 'a91c7b26-f710-4ef4-b0d8-a21c4e12b982' has exited > I0729 17:10:21.470084 1213 slave.cpp:2528] Executor '1' of framework > 20140729-171010-16842879-54701-1176-0000 terminated with signal Killed > I0729 17:10:21.470188 1213 slave.cpp:2662] Cleaning up executor '1' of > framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:21.470685 1213 slave.cpp:2737] Cleaning up framework > 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:21.471007 1213 gc.cpp:56] Scheduling > '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/slaves/20140729-171010-16842879-54701-1176-0/frameworks/20140729-171010-16842879-54701-1176-0000/executors/1/runs/a91c7b26-f710-4ef4-b0d8-a21c4e12b982' > for gc 6.99999455449185days in the future > I0729 17:10:21.471166 1213 gc.cpp:56] Scheduling > '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/slaves/20140729-171010-16842879-54701-1176-0/frameworks/20140729-171010-16842879-54701-1176-0000/executors/1' > for gc 6.99999455256889days in the future > I0729 17:10:21.471251 1213 gc.cpp:56] Scheduling > '/tmp/HealthCheckTest_GracePeriod_f0DeJJ/slaves/20140729-171010-16842879-54701-1176-0/frameworks/20140729-171010-16842879-54701-1176-0000' > for gc 6.99999455120296days in the future > I0729 17:10:21.471336 1213 status_update_manager.cpp:282] Closing status > update streams for framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:21.471395 1213 status_update_manager.cpp:530] Cleaning up status > update stream for task 1 of framework 20140729-171010-16842879-54701-1176-0000 > I0729 17:10:21.472653 1176 slave.cpp:485] Slave terminating > [ FAILED ] HealthCheckTest.GracePeriod (11019 ms) > {noformat} -- This message was sent by Atlassian JIRA (v6.2#6252)