Build failed in Jenkins: mesos-reviewbot #9800

2015-11-20 Thread Apache Jenkins Server
See 

--
[...truncated 7908 lines...]
rm -f v1/.dirstamp
rm -f version/.deps/.dirstamp
rm -f version/.dirstamp
rm -f watcher/.deps/.dirstamp
rm -f watcher/.dirstamp
rm -rf slave/containerizer/.libs slave/containerizer/_libs
rm -f zookeeper/.deps/.dirstamp
rm -f zookeeper/.dirstamp
rm -rf slave/containerizer/mesos/.libs slave/containerizer/mesos/_libs
rm -rf slave/containerizer/mesos/isolators/cgroups/.libs 
slave/containerizer/mesos/isolators/cgroups/_libs
rm -rf slave/containerizer/mesos/isolators/filesystem/.libs 
slave/containerizer/mesos/isolators/filesystem/_libs
rm -rf slave/containerizer/mesos/isolators/namespaces/.libs 
slave/containerizer/mesos/isolators/namespaces/_libs
rm -rf slave/containerizer/mesos/isolators/network/.libs 
slave/containerizer/mesos/isolators/network/_libs
rm -rf slave/containerizer/mesos/isolators/posix/.libs 
slave/containerizer/mesos/isolators/posix/_libs
rm -rf slave/containerizer/mesos/provisioner/.libs 
slave/containerizer/mesos/provisioner/_libs
rm -rf slave/containerizer/mesos/provisioner/appc/.libs 
slave/containerizer/mesos/provisioner/appc/_libs
rm -rf slave/containerizer/mesos/provisioner/backends/.libs 
slave/containerizer/mesos/provisioner/backends/_libs
rm -rf slave/containerizer/mesos/provisioner/docker/.libs 
slave/containerizer/mesos/provisioner/docker/_libs
rm -rf slave/qos_controllers/.libs slave/qos_controllers/_libs
rm -rf slave/resource_estimators/.libs slave/resource_estimators/_libs
rm -rf state/.libs state/_libs
rm -rf usage/.libs usage/_libs
rm -rf v1/.libs v1/_libs
rm -rf version/.libs version/_libs
rm -rf watcher/.libs watcher/_libs
rm -rf zookeeper/.libs zookeeper/_libs
rm -f tests/common/*.o
rm -f tests/containerizer/*.o
rm -f usage/*.o
rm -f usage/*.lo
rm -f v1/*.o
rm -f v1/*.lo
rm -f version/*.o
rm -f version/*.lo
rm -f watcher/*.o
rm -f watcher/*.lo
rm -f zookeeper/*.o
rm -f zookeeper/*.lo
rm -rf ../include/mesos/.deps ../include/mesos/authentication/.deps 
../include/mesos/authorizer/.deps ../include/mesos/containerizer/.deps 
../include/mesos/executor/.deps ../include/mesos/fetcher/.deps 
../include/mesos/maintenance/.deps ../include/mesos/master/.deps 
../include/mesos/module/.deps ../include/mesos/quota/.deps 
../include/mesos/scheduler/.deps ../include/mesos/slave/.deps 
../include/mesos/v1/.deps ../include/mesos/v1/executor/.deps 
../include/mesos/v1/scheduler/.deps authentication/cram_md5/.deps 
authorizer/.deps authorizer/local/.deps cli/.deps common/.deps docker/.deps 
examples/.deps exec/.deps files/.deps health-check/.deps hook/.deps 
internal/.deps java/jni/.deps jvm/.deps jvm/org/apache/.deps launcher/.deps 
linux/.deps linux/routing/.deps linux/routing/diagnosis/.deps 
linux/routing/filter/.deps linux/routing/link/.deps 
linux/routing/queueing/.deps local/.deps log/.deps log/tool/.deps logging/.deps 
master/.deps master/allocator/.deps master/allocator/mesos/.deps 
master/allocator/sorter/drf/.deps messages/.deps module/.deps sched/.deps 
scheduler/.deps slave/.deps slave/containerizer/.deps 
slave/containerizer/mesos/.deps 
slave/containerizer/mesos/isolators/cgroups/.deps 
slave/containerizer/mesos/isolators/filesystem/.deps 
slave/containerizer/mesos/isolators/namespaces/.deps 
slave/containerizer/mesos/isolators/network/.deps 
slave/containerizer/mesos/isolators/posix/.deps 
slave/containerizer/mesos/provisioner/.deps 
slave/containerizer/mesos/provisioner/appc/.deps 
slave/containerizer/mesos/provisioner/backends/.deps 
slave/containerizer/mesos/provisioner/docker/.deps slave/qos_controllers/.deps 
slave/resource_estimators/.deps state/.deps tests/.deps tests/common/.deps 
tests/containerizer/.deps usage/.deps v1/.deps version/.deps watcher/.deps 
zookeeper/.deps
rm -f Makefile
make[2]: Leaving directory 
`
rm -f config.status config.cache config.log configure.lineno 
config.status.lineno
rm -f Makefile
make[1]: Leaving directory 
`
if test -d "mesos-0.26.0"; then find "mesos-0.26.0" -type d ! -perm -200 -exec 
chmod u+w {} ';' && rm -rf "mesos-0.26.0" || { sleep 5 && rm -rf 
"mesos-0.26.0"; }; else :; fi
==
mesos-0.26.0 archives ready for distribution: 
mesos-0.26.0.tar.gz
==

real22m33.719s
user112m28.625s
sys 8m31.132s
+ chmod -R +w 3rdparty aclocal.m4 ar-lib autom4te.cache bin bootstrap CHANGELOG 
cmake CMakeLists.txt compile config.guess config.log config.lt config.status 
config.sub configure configure.ac depcomp Dockerfile docs Doxyfile include 
install-sh libtool LICENSE ltmain.sh m4 Makefile Makefile.am Makefile.in 
mesos-0.26.0.tar.gz mesos.pc mesos.pc.in missing mpi NOTICE README.md site src 
support
+ git clean -fdx
Removing .clang-format
Removing .gitignore
Removing .libs/

Jenkins build is back to normal : Mesos » gcc,--verbose,ubuntu:14.04,docker||Hadoop #1250

2015-11-20 Thread Apache Jenkins Server
See 




Build failed in Jenkins: Mesos » gcc,--verbose,ubuntu:14.04,docker||Hadoop #1249

2015-11-20 Thread Apache Jenkins Server
See 


Changes:

[benjamin.mahler] Restored the ABORT message in Subprocess to show the execvpe 
path.

--
[...truncated 140216 lines...]
I1120 12:08:25.244694 29833 slave.cpp:2009] Asked to shut down framework 
2319667c-dde4-4f05-b5bd-8d9dce373512- by @0.0.0.0:0
I1120 12:08:25.244734 29833 slave.cpp:2034] Shutting down framework 
2319667c-dde4-4f05-b5bd-8d9dce373512-
I1120 12:08:25.244807 29833 slave.cpp:3657] Cleaning up executor 'default' of 
framework 2319667c-dde4-4f05-b5bd-8d9dce373512- at 
executor(132)@172.17.5.188:41228
I1120 12:08:25.245301 29838 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_j02Pv2/slaves/2319667c-dde4-4f05-b5bd-8d9dce373512-S0/frameworks/2319667c-dde4-4f05-b5bd-8d9dce373512-/executors/default/runs/12eebe8b-4a1f-4516-a73c-8e68a526c6d8'
 for gc 6.9716431704days in the future
I1120 12:08:25.245405 29833 slave.cpp:3745] Cleaning up framework 
2319667c-dde4-4f05-b5bd-8d9dce373512-
I1120 12:08:25.245801 29838 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_j02Pv2/slaves/2319667c-dde4-4f05-b5bd-8d9dce373512-S0/frameworks/2319667c-dde4-4f05-b5bd-8d9dce373512-/executors/default'
 for gc 6.9716053037days in the future
I1120 12:08:25.245858 29833 status_update_manager.cpp:284] Closing status 
update streams for framework 2319667c-dde4-4f05-b5bd-8d9dce373512-
I1120 12:08:25.246083 29833 status_update_manager.cpp:530] Cleaning up status 
update stream for task e046ed13-d6f5-4ea1-ba62-99337755a5ad of framework 
2319667c-dde4-4f05-b5bd-8d9dce373512-
I1120 12:08:25.246162 29838 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_j02Pv2/slaves/2319667c-dde4-4f05-b5bd-8d9dce373512-S0/frameworks/2319667c-dde4-4f05-b5bd-8d9dce373512-'
 for gc 6.9715612148days in the future
[   OK ] ContentType/SchedulerTest.Message/1 (123 ms)
[ RUN  ] ContentType/SchedulerTest.Request/0
I1120 12:08:25.261420 29806 leveldb.cpp:176] Opened db in 3.502393ms
I1120 12:08:25.262426 29806 leveldb.cpp:183] Compacted db in 939509ns
I1120 12:08:25.262498 29806 leveldb.cpp:198] Created db iterator in 24099ns
I1120 12:08:25.262514 29806 leveldb.cpp:204] Seeked to beginning of db in 1985ns
I1120 12:08:25.262526 29806 leveldb.cpp:273] Iterated through 0 keys in the db 
in 403ns
I1120 12:08:25.262579 29806 replica.cpp:780] Replica recovered with log 
positions 0 -> 0 with 1 holes and 0 unlearned
I1120 12:08:25.264700 29825 recover.cpp:449] Starting replica recovery
I1120 12:08:25.265362 29838 recover.cpp:475] Replica is in EMPTY status
I1120 12:08:25.266561 29839 master.cpp:367] Master 
0fdff1f7-cb45-4d58-b44c-cba92980cf0f (a51e6bb03b55) started on 
172.17.5.188:41228
I1120 12:08:25.266968 29839 master.cpp:369] Flags at startup: --acls="" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate="false" --authenticate_slaves="true" --authenticators="crammd5" 
--authorizers="local" --credentials="/tmp/n5mcWN/credentials" 
--framework_sorter="drf" --help="false" --hostname_lookup="true" 
--initialize_driver_logging="true" --log_auto_initialize="true" 
--logbufsecs="0" --logging_level="INFO" --max_slave_ping_timeouts="5" 
--quiet="false" --recovery_slave_removal_limit="100%" 
--registry="replicated_log" --registry_fetch_timeout="1mins" 
--registry_store_timeout="25secs" --registry_strict="true" 
--root_submissions="true" --slave_ping_timeout="15secs" 
--slave_reregister_timeout="10mins" --user_sorter="drf" --version="false" 
--webui_dir="/mesos/mesos-0.26.0/_inst/share/mesos/webui" 
--work_dir="/tmp/n5mcWN/master" --zk_session_timeout="10secs"
I1120 12:08:25.267544 29839 master.cpp:416] Master allowing unauthenticated 
frameworks to register
I1120 12:08:25.267655 29839 master.cpp:419] Master only allowing authenticated 
slaves to register
I1120 12:08:25.267861 29839 credentials.hpp:37] Loading credentials for 
authentication from '/tmp/n5mcWN/credentials'
I1120 12:08:25.268374 29839 master.cpp:458] Using default 'crammd5' 
authenticator
I1120 12:08:25.268813 29839 master.cpp:495] Authorization enabled
I1120 12:08:25.270114 29831 hierarchical.cpp:153] Initialized hierarchical 
allocator process
I1120 12:08:25.270238 29832 whitelist_watcher.cpp:79] No whitelist given
I1120 12:08:25.273159 29830 replica.cpp:676] Replica in EMPTY status received a 
broadcasted recover request from (10652)@172.17.5.188:41228
I1120 12:08:25.274507 29826 master.cpp:1612] The newly elected leader is 
master@172.17.5.188:41228 with id 0fdff1f7-cb45-4d58-b44c-cba92980cf0f
I1120 12:08:25.274543 29826 master.cpp:1625] Elected as the leading master!
I1120 12:08:25.274565 29826 master.cpp:1385] Recovering from registrar
I1120 12:08:25.274737 29832 registrar.cpp:309] Recovering registrar
I1120 12:08:25.275318 29825 recover.cpp:195] Received a recover response from a 

Re: Build failed in Jenkins: mesos-reviewbot #9787

2015-11-20 Thread Benjamin Mahler
Hey Vinod / Jie,

It appears that these recent failures are from the review here
https://reviews.apache.org/r/40418/. It looks like because
verify_reviews.py hits a "500 Internal Server Error" it barfs out and
emails the list. Unfortunately, this ends up being a bit misleading in this
case because it looks at first glance like the build is failing (I was
tripped up by the CurlFetcherPluginTest failures above).

Any ideas on how to make the reviewbot failures less misleading?

On Thu, Nov 19, 2015 at 11:52 PM, Apache Jenkins Server <
jenk...@builds.apache.org> wrote:

> See 
>
> --
> [...truncated 148679 lines...]
> I1119 22:51:39.306417 30230 replica.cpp:323] Persisted replica status to
> VOTING
> I1119 22:51:39.306602 30229 recover.cpp:580] Successfully joined the Paxos
> group
> I1119 22:51:39.306910 30229 recover.cpp:464] Recover process terminated
> I1119 22:51:39.307488 30231 log.cpp:661] Attempting to start the writer
> I1119 22:51:39.309092 30227 replica.cpp:496] Replica received implicit
> promise request from (10553)@172.17.19.74:52056 with proposal 1
> I1119 22:51:39.309515 30227 leveldb.cpp:306] Persisting metadata (8 bytes)
> to leveldb took 376214ns
> I1119 22:51:39.309556 30227 replica.cpp:345] Persisted promised to 1
> I1119 22:51:39.310180 30221 coordinator.cpp:240] Coordinator attempting to
> fill missing positions
> I1119 22:51:39.311311 30224 replica.cpp:391] Replica received explicit
> promise request from (10554)@172.17.19.74:52056 for position 0 with
> proposal 2
> I1119 22:51:39.311645 30224 leveldb.cpp:343] Persisting action (8 bytes)
> to leveldb took 296670ns
> I1119 22:51:39.311667 30224 replica.cpp:715] Persisted action at 0
> I1119 22:51:39.312527 30231 replica.cpp:540] Replica received write
> request for position 0 from (10555)@172.17.19.74:52056
> I1119 22:51:39.312585 30231 leveldb.cpp:438] Reading position from leveldb
> took 26213ns
> I1119 22:51:39.312921 30231 leveldb.cpp:343] Persisting action (14 bytes)
> to leveldb took 288780ns
> I1119 22:51:39.312943 30231 replica.cpp:715] Persisted action at 0
> I1119 22:51:39.313567 30220 replica.cpp:694] Replica received learned
> notice for position 0 from @0.0.0.0:0
> I1119 22:51:39.313930 30220 leveldb.cpp:343] Persisting action (16 bytes)
> to leveldb took 332743ns
> I1119 22:51:39.313953 30220 replica.cpp:715] Persisted action at 0
> I1119 22:51:39.313969 30220 replica.cpp:700] Replica learned NOP action at
> position 0
> I1119 22:51:39.314502 30224 log.cpp:677] Writer started with ending
> position 0
> I1119 22:51:39.315490 30220 leveldb.cpp:438] Reading position from leveldb
> took 24579ns
> I1119 22:51:39.316335 30229 registrar.cpp:342] Successfully fetched the
> registry (0B) in 10.753792ms
> I1119 22:51:39.316457 30229 registrar.cpp:441] Applied 1 operations in
> 27472ns; attempting to update the 'registry'
> I1119 22:51:39.317178 30222 log.cpp:685] Attempting to append 176 bytes to
> the log
> I1119 22:51:39.317297 30223 coordinator.cpp:350] Coordinator attempting to
> write APPEND action at position 1
> I1119 22:51:39.318049 30219 replica.cpp:540] Replica received write
> request for position 1 from (10556)@172.17.19.74:52056
> I1119 22:51:39.318389 30219 leveldb.cpp:343] Persisting action (195 bytes)
> to leveldb took 300611ns
> I1119 22:51:39.318411 30219 replica.cpp:715] Persisted action at 1
> I1119 22:51:39.318946 30222 replica.cpp:694] Replica received learned
> notice for position 1 from @0.0.0.0:0
> I1119 22:51:39.319262 30222 leveldb.cpp:343] Persisting action (197 bytes)
> to leveldb took 282235ns
> I1119 22:51:39.319286 30222 replica.cpp:715] Persisted action at 1
> I1119 22:51:39.319311 30222 replica.cpp:700] Replica learned APPEND action
> at position 1
> I1119 22:51:39.320333 30226 registrar.cpp:486] Successfully updated the
> 'registry' in 3.81696ms
> I1119 22:51:39.320457 30226 registrar.cpp:372] Successfully recovered
> registrar
> I1119 22:51:39.320579 30220 log.cpp:704] Attempting to truncate the log to
> 1
> I1119 22:51:39.320813 30226 coordinator.cpp:350] Coordinator attempting to
> write TRUNCATE action at position 2
> I1119 22:51:39.320878 30227 master.cpp:1422] Recovered 0 slaves from the
> Registry (137B) ; allowing 10mins for slaves to re-register
> I1119 22:51:39.321463 30233 replica.cpp:540] Replica received write
> request for position 2 from (10557)@172.17.19.74:52056
> I1119 22:51:39.321768 30233 leveldb.cpp:343] Persisting action (16 bytes)
> to leveldb took 256133ns
> I1119 22:51:39.321789 30233 replica.cpp:715] Persisted action at 2
> I1119 22:51:39.322340 30226 replica.cpp:694] Replica received learned
> notice for position 2 from @0.0.0.0:0
> I1119 22:51:39.322696 30226 leveldb.cpp:343] Persisting action (18 bytes)
> to leveldb took 325723ns
> I1119 22:51:39.322742 30226 leveldb.cpp:401] Deleting ~1 keys from leveldb
> took 24686ns
> I1119 22:51:39.322764 30226 replica.cpp:715] Persisted 

Build failed in Jenkins: Mesos » clang,--verbose --enable-libevent --enable-ssl,ubuntu:14.04,docker||Hadoop #1254

2015-11-20 Thread Apache Jenkins Server
See 


Changes:

[vinodkone] Updated ReviewBot to handle URL errors.

--
[...truncated 142291 lines...]
I1120 21:51:12.493969 29722 slave.cpp:3657] Cleaning up executor 'default' of 
framework 93061a2a-f83d-482f-a541-210eba715239- at 
executor(132)@172.17.19.237:37352
I1120 21:51:12.494138 29727 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_bxOKNB/slaves/93061a2a-f83d-482f-a541-210eba715239-S0/frameworks/93061a2a-f83d-482f-a541-210eba715239-/executors/default/runs/5c6f33c9-0285-4197-bb8a-e6a7d1edf1c2'
 for gc 6.9428171259days in the future
I1120 21:51:12.494225 29722 slave.cpp:3745] Cleaning up framework 
93061a2a-f83d-482f-a541-210eba715239-
I1120 21:51:12.494254 29727 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_bxOKNB/slaves/93061a2a-f83d-482f-a541-210eba715239-S0/frameworks/93061a2a-f83d-482f-a541-210eba715239-/executors/default'
 for gc 6.9428027852days in the future
I1120 21:51:12.494326 29732 status_update_manager.cpp:284] Closing status 
update streams for framework 93061a2a-f83d-482f-a541-210eba715239-
I1120 21:51:12.494374 29727 gc.cpp:56] Scheduling 
'/tmp/ContentType_SchedulerTest_Message_1_bxOKNB/slaves/93061a2a-f83d-482f-a541-210eba715239-S0/frameworks/93061a2a-f83d-482f-a541-210eba715239-'
 for gc 6.9427868148days in the future
I1120 21:51:12.494380 29732 status_update_manager.cpp:530] Cleaning up status 
update stream for task 60715210-ad6c-41d1-b532-8960accd4036 of framework 
93061a2a-f83d-482f-a541-210eba715239-
[   OK ] ContentType/SchedulerTest.Message/1 (752 ms)
[ RUN  ] ContentType/SchedulerTest.Request/0
I1120 21:51:12.610169 29699 leveldb.cpp:176] Opened db in 111.273479ms
I1120 21:51:12.658470 29699 leveldb.cpp:183] Compacted db in 48.2756ms
I1120 21:51:12.658529 29699 leveldb.cpp:198] Created db iterator in 29234ns
I1120 21:51:12.658546 29699 leveldb.cpp:204] Seeked to beginning of db in 7352ns
I1120 21:51:12.658557 29699 leveldb.cpp:273] Iterated through 0 keys in the db 
in 5807ns
I1120 21:51:12.658598 29699 replica.cpp:780] Replica recovered with log 
positions 0 -> 0 with 1 holes and 0 unlearned
I1120 21:51:12.659147 29725 recover.cpp:449] Starting replica recovery
I1120 21:51:12.659356 29725 recover.cpp:475] Replica is in EMPTY status
I1120 21:51:12.660328 29718 replica.cpp:676] Replica in EMPTY status received a 
broadcasted recover request from (10585)@172.17.19.237:37352
I1120 21:51:12.660915 29721 recover.cpp:195] Received a recover response from a 
replica in EMPTY status
I1120 21:51:12.661278 29731 recover.cpp:566] Updating replica status to STARTING
I1120 21:51:12.661785 29725 master.cpp:367] Master 
fda0487f-9184-412a-b1f3-10b04b13d711 (d547b5872019) started on 
172.17.19.237:37352
I1120 21:51:12.661821 29725 master.cpp:369] Flags at startup: --acls="" 
--allocation_interval="1secs" --allocator="HierarchicalDRF" 
--authenticate="false" --authenticate_slaves="true" --authenticators="crammd5" 
--authorizers="local" --credentials="/tmp/9Zkxfs/credentials" 
--framework_sorter="drf" --help="false" --hostname_lookup="true" 
--initialize_driver_logging="true" --log_auto_initialize="true" 
--logbufsecs="0" --logging_level="INFO" --max_slave_ping_timeouts="5" 
--quiet="false" --recovery_slave_removal_limit="100%" 
--registry="replicated_log" --registry_fetch_timeout="1mins" 
--registry_store_timeout="25secs" --registry_strict="true" 
--root_submissions="true" --slave_ping_timeout="15secs" 
--slave_reregister_timeout="10mins" --user_sorter="drf" --version="false" 
--webui_dir="/mesos/mesos-0.26.0/_inst/share/mesos/webui" 
--work_dir="/tmp/9Zkxfs/master" --zk_session_timeout="10secs"
I1120 21:51:12.662086 29725 master.cpp:416] Master allowing unauthenticated 
frameworks to register
I1120 21:51:12.662103 29725 master.cpp:419] Master only allowing authenticated 
slaves to register
I1120 21:51:12.662117 29725 credentials.hpp:37] Loading credentials for 
authentication from '/tmp/9Zkxfs/credentials'
I1120 21:51:12.662438 29725 master.cpp:458] Using default 'crammd5' 
authenticator
I1120 21:51:12.662575 29725 master.cpp:495] Authorization enabled
I1120 21:51:12.662950 29730 hierarchical.cpp:153] Initialized hierarchical 
allocator process
I1120 21:51:12.662971 29724 whitelist_watcher.cpp:79] No whitelist given
I1120 21:51:12.664731 29730 master.cpp:1612] The newly elected leader is 
master@172.17.19.237:37352 with id fda0487f-9184-412a-b1f3-10b04b13d711
I1120 21:51:12.664777 29730 master.cpp:1625] Elected as the leading master!
I1120 21:51:12.664801 29730 master.cpp:1385] Recovering from registrar
I1120 21:51:12.664940 29719 registrar.cpp:309] Recovering registrar
I1120 21:51:12.692426 29721 leveldb.cpp:306] Persisting metadata (8 bytes) to 
leveldb took 31.057348ms
I1120 21:51:12.692474 

Re: Build failed in Jenkins: mesos-reviewbot #9787

2015-11-20 Thread Vinod Kone
Pushed a fix to gracefully handle URL errors. This should avoid the
confusion.

On Fri, Nov 20, 2015 at 7:45 AM, Benjamin Mahler 
wrote:

> Hey Vinod / Jie,
>
> It appears that these recent failures are from the review here
> https://reviews.apache.org/r/40418/. It looks like because
> verify_reviews.py hits a "500 Internal Server Error" it barfs out and
> emails the list. Unfortunately, this ends up being a bit misleading in this
> case because it looks at first glance like the build is failing (I was
> tripped up by the CurlFetcherPluginTest failures above).
>
> Any ideas on how to make the reviewbot failures less misleading?
>
> On Thu, Nov 19, 2015 at 11:52 PM, Apache Jenkins Server <
> jenk...@builds.apache.org> wrote:
>
>> See 
>>
>> --
>> [...truncated 148679 lines...]
>> I1119 22:51:39.306417 30230 replica.cpp:323] Persisted replica status to
>> VOTING
>> I1119 22:51:39.306602 30229 recover.cpp:580] Successfully joined the
>> Paxos group
>> I1119 22:51:39.306910 30229 recover.cpp:464] Recover process terminated
>> I1119 22:51:39.307488 30231 log.cpp:661] Attempting to start the writer
>> I1119 22:51:39.309092 30227 replica.cpp:496] Replica received implicit
>> promise request from (10553)@172.17.19.74:52056 with proposal 1
>> I1119 22:51:39.309515 30227 leveldb.cpp:306] Persisting metadata (8
>> bytes) to leveldb took 376214ns
>> I1119 22:51:39.309556 30227 replica.cpp:345] Persisted promised to 1
>> I1119 22:51:39.310180 30221 coordinator.cpp:240] Coordinator attempting
>> to fill missing positions
>> I1119 22:51:39.311311 30224 replica.cpp:391] Replica received explicit
>> promise request from (10554)@172.17.19.74:52056 for position 0 with
>> proposal 2
>> I1119 22:51:39.311645 30224 leveldb.cpp:343] Persisting action (8 bytes)
>> to leveldb took 296670ns
>> I1119 22:51:39.311667 30224 replica.cpp:715] Persisted action at 0
>> I1119 22:51:39.312527 30231 replica.cpp:540] Replica received write
>> request for position 0 from (10555)@172.17.19.74:52056
>> I1119 22:51:39.312585 30231 leveldb.cpp:438] Reading position from
>> leveldb took 26213ns
>> I1119 22:51:39.312921 30231 leveldb.cpp:343] Persisting action (14 bytes)
>> to leveldb took 288780ns
>> I1119 22:51:39.312943 30231 replica.cpp:715] Persisted action at 0
>> I1119 22:51:39.313567 30220 replica.cpp:694] Replica received learned
>> notice for position 0 from @0.0.0.0:0
>> I1119 22:51:39.313930 30220 leveldb.cpp:343] Persisting action (16 bytes)
>> to leveldb took 332743ns
>> I1119 22:51:39.313953 30220 replica.cpp:715] Persisted action at 0
>> I1119 22:51:39.313969 30220 replica.cpp:700] Replica learned NOP action
>> at position 0
>> I1119 22:51:39.314502 30224 log.cpp:677] Writer started with ending
>> position 0
>> I1119 22:51:39.315490 30220 leveldb.cpp:438] Reading position from
>> leveldb took 24579ns
>> I1119 22:51:39.316335 30229 registrar.cpp:342] Successfully fetched the
>> registry (0B) in 10.753792ms
>> I1119 22:51:39.316457 30229 registrar.cpp:441] Applied 1 operations in
>> 27472ns; attempting to update the 'registry'
>> I1119 22:51:39.317178 30222 log.cpp:685] Attempting to append 176 bytes
>> to the log
>> I1119 22:51:39.317297 30223 coordinator.cpp:350] Coordinator attempting
>> to write APPEND action at position 1
>> I1119 22:51:39.318049 30219 replica.cpp:540] Replica received write
>> request for position 1 from (10556)@172.17.19.74:52056
>> I1119 22:51:39.318389 30219 leveldb.cpp:343] Persisting action (195
>> bytes) to leveldb took 300611ns
>> I1119 22:51:39.318411 30219 replica.cpp:715] Persisted action at 1
>> I1119 22:51:39.318946 30222 replica.cpp:694] Replica received learned
>> notice for position 1 from @0.0.0.0:0
>> I1119 22:51:39.319262 30222 leveldb.cpp:343] Persisting action (197
>> bytes) to leveldb took 282235ns
>> I1119 22:51:39.319286 30222 replica.cpp:715] Persisted action at 1
>> I1119 22:51:39.319311 30222 replica.cpp:700] Replica learned APPEND
>> action at position 1
>> I1119 22:51:39.320333 30226 registrar.cpp:486] Successfully updated the
>> 'registry' in 3.81696ms
>> I1119 22:51:39.320457 30226 registrar.cpp:372] Successfully recovered
>> registrar
>> I1119 22:51:39.320579 30220 log.cpp:704] Attempting to truncate the log
>> to 1
>> I1119 22:51:39.320813 30226 coordinator.cpp:350] Coordinator attempting
>> to write TRUNCATE action at position 2
>> I1119 22:51:39.320878 30227 master.cpp:1422] Recovered 0 slaves from the
>> Registry (137B) ; allowing 10mins for slaves to re-register
>> I1119 22:51:39.321463 30233 replica.cpp:540] Replica received write
>> request for position 2 from (10557)@172.17.19.74:52056
>> I1119 22:51:39.321768 30233 leveldb.cpp:343] Persisting action (16 bytes)
>> to leveldb took 256133ns
>> I1119 22:51:39.321789 30233 replica.cpp:715] Persisted action at 2
>> I1119 22:51:39.322340 30226 replica.cpp:694] Replica received learned
>> notice for position 2 

Jenkins build is back to normal : Mesos » clang,--verbose --enable-libevent --enable-ssl,ubuntu:14.04,docker||Hadoop #1255

2015-11-20 Thread Apache Jenkins Server
See 




Build failed in Jenkins: mesos-reviewbot #9809

2015-11-20 Thread Apache Jenkins Server
See 

Changes:

[vinodkone] Updated ReviewBot to handle URL errors.

--
[...truncated 7919 lines...]
rm -rf master/allocator/mesos/.libs master/allocator/mesos/_libs
rm -rf master/allocator/sorter/drf/.libs master/allocator/sorter/drf/_libs
rm -rf messages/.libs messages/_libs
rm -rf module/.libs module/_libs
rm -rf sched/.libs sched/_libs
rm -rf scheduler/.libs scheduler/_libs
rm -rf slave/.libs slave/_libs
rm -rf slave/containerizer/.libs slave/containerizer/_libs
rm -rf slave/containerizer/mesos/.libs slave/containerizer/mesos/_libs
rm -rf slave/containerizer/mesos/isolators/cgroups/.libs 
slave/containerizer/mesos/isolators/cgroups/_libs
rm -rf slave/containerizer/mesos/isolators/filesystem/.libs 
slave/containerizer/mesos/isolators/filesystem/_libs
rm -rf slave/containerizer/mesos/isolators/namespaces/.libs 
slave/containerizer/mesos/isolators/namespaces/_libs
rm -rf slave/containerizer/mesos/isolators/network/.libs 
slave/containerizer/mesos/isolators/network/_libs
rm -rf slave/containerizer/mesos/isolators/posix/.libs 
slave/containerizer/mesos/isolators/posix/_libs
rm -rf slave/containerizer/mesos/provisioner/.libs 
slave/containerizer/mesos/provisioner/_libs
rm -f tests/common/*.o
rm -f tests/containerizer/*.o
rm -rf slave/containerizer/mesos/provisioner/appc/.libs 
slave/containerizer/mesos/provisioner/appc/_libs
rm -rf slave/containerizer/mesos/provisioner/backends/.libs 
slave/containerizer/mesos/provisioner/backends/_libs
rm -rf slave/containerizer/mesos/provisioner/docker/.libs 
slave/containerizer/mesos/provisioner/docker/_libs
rm -f usage/*.o
rm -rf slave/qos_controllers/.libs slave/qos_controllers/_libs
rm -rf slave/resource_estimators/.libs slave/resource_estimators/_libs
rm -f usage/*.lo
rm -f v1/*.o
rm -rf state/.libs state/_libs
rm -f v1/*.lo
rm -f version/*.o
rm -rf usage/.libs usage/_libs
rm -f version/*.lo
rm -f watcher/*.o
rm -rf v1/.libs v1/_libs
rm -f watcher/*.lo
rm -rf version/.libs version/_libs
rm -f zookeeper/*.o
rm -rf watcher/.libs watcher/_libs
rm -f zookeeper/*.lo
rm -rf zookeeper/.libs zookeeper/_libs
rm -rf ../include/mesos/.deps ../include/mesos/authentication/.deps 
../include/mesos/authorizer/.deps ../include/mesos/containerizer/.deps 
../include/mesos/executor/.deps ../include/mesos/fetcher/.deps 
../include/mesos/maintenance/.deps ../include/mesos/master/.deps 
../include/mesos/module/.deps ../include/mesos/quota/.deps 
../include/mesos/scheduler/.deps ../include/mesos/slave/.deps 
../include/mesos/v1/.deps ../include/mesos/v1/executor/.deps 
../include/mesos/v1/scheduler/.deps authentication/cram_md5/.deps 
authorizer/.deps authorizer/local/.deps cli/.deps common/.deps docker/.deps 
examples/.deps exec/.deps files/.deps health-check/.deps hook/.deps 
internal/.deps java/jni/.deps jvm/.deps jvm/org/apache/.deps launcher/.deps 
linux/.deps linux/routing/.deps linux/routing/diagnosis/.deps 
linux/routing/filter/.deps linux/routing/link/.deps 
linux/routing/queueing/.deps local/.deps log/.deps log/tool/.deps logging/.deps 
master/.deps master/allocator/.deps master/allocator/mesos/.deps 
master/allocator/sorter/drf/.deps messages/.deps module/.deps sched/.deps 
scheduler/.deps slave/.deps slave/containerizer/.deps 
slave/containerizer/mesos/.deps 
slave/containerizer/mesos/isolators/cgroups/.deps 
slave/containerizer/mesos/isolators/filesystem/.deps 
slave/containerizer/mesos/isolators/namespaces/.deps 
slave/containerizer/mesos/isolators/network/.deps 
slave/containerizer/mesos/isolators/posix/.deps 
slave/containerizer/mesos/provisioner/.deps 
slave/containerizer/mesos/provisioner/appc/.deps 
slave/containerizer/mesos/provisioner/backends/.deps 
slave/containerizer/mesos/provisioner/docker/.deps slave/qos_controllers/.deps 
slave/resource_estimators/.deps state/.deps tests/.deps tests/common/.deps 
tests/containerizer/.deps usage/.deps v1/.deps version/.deps watcher/.deps 
zookeeper/.deps
rm -f Makefile
make[2]: Leaving directory 
`
rm -f config.status config.cache config.log configure.lineno 
config.status.lineno
rm -f Makefile
make[1]: Leaving directory 
`
if test -d "mesos-0.26.0"; then find "mesos-0.26.0" -type d ! -perm -200 -exec 
chmod u+w {} ';' && rm -rf "mesos-0.26.0" || { sleep 5 && rm -rf 
"mesos-0.26.0"; }; else :; fi
==
mesos-0.26.0 archives ready for distribution: 
mesos-0.26.0.tar.gz
==

real24m55.824s
user113m34.676s
sys 9m22.944s
+ chmod -R +w 3rdparty aclocal.m4 ar-lib autom4te.cache bin bootstrap CHANGELOG 
cmake CMakeLists.txt compile config.guess config.log config.lt config.status 
config.sub configure configure.ac depcomp Dockerfile docs Doxyfile include 
install-sh libtool LICENSE ltmain.sh m4 

Jenkins build is back to normal : mesos-reviewbot #9811

2015-11-20 Thread Apache Jenkins Server
See 



Build failed in Jenkins: mesos-reviewbot #9810

2015-11-20 Thread Apache Jenkins Server
See 

Changes:

[yujie.jay] Added URI protobuf definition.

[yujie.jay] Added URI fetcher interface.

[yujie.jay] Added streaming and construction methods for URI.

[yujie.jay] Added curl based URI fetcher plugin.

[yujie.jay] Added a test filter for CURL tests.

[yujie.jay] Changed HDFS wrapper from a struct to a class.

[yujie.jay] Fixed the license header in hdfs.hpp.

[yujie.jay] Moved HDFS wrapper implementation to a cpp file.

[yujie.jay] Fixed a few style issues in HDFS wrapper code.

[yujie.jay] Used factory method to create HDFS client.

[tnachen] Added containerizer option to mesos-execute.

[yujie.jay] Always create non-IP egress filters in port mapping isolator.

[tnachen] Added backtick usage in comments to the C++ style guide.

[tnachen] Added logging untar process error in docker store.

--
[...truncated 7986 lines...]
rm -rf slave/containerizer/mesos/isolators/cgroups/.libs 
slave/containerizer/mesos/isolators/cgroups/_libs
rm -f tests/containerizer/*.o
rm -rf slave/containerizer/mesos/isolators/filesystem/.libs 
slave/containerizer/mesos/isolators/filesystem/_libs
rm -rf slave/containerizer/mesos/isolators/namespaces/.libs 
slave/containerizer/mesos/isolators/namespaces/_libs
rm -rf slave/containerizer/mesos/isolators/network/.libs 
slave/containerizer/mesos/isolators/network/_libs
rm -rf slave/containerizer/mesos/isolators/posix/.libs 
slave/containerizer/mesos/isolators/posix/_libs
rm -rf slave/containerizer/mesos/provisioner/.libs 
slave/containerizer/mesos/provisioner/_libs
rm -rf slave/containerizer/mesos/provisioner/appc/.libs 
slave/containerizer/mesos/provisioner/appc/_libs
rm -rf slave/containerizer/mesos/provisioner/backends/.libs 
slave/containerizer/mesos/provisioner/backends/_libs
rm -rf slave/containerizer/mesos/provisioner/docker/.libs 
slave/containerizer/mesos/provisioner/docker/_libs
rm -f uri/*.o
rm -rf slave/qos_controllers/.libs slave/qos_controllers/_libs
rm -f uri/*.lo
rm -rf slave/resource_estimators/.libs slave/resource_estimators/_libs
rm -f uri/fetchers/*.o
rm -rf state/.libs state/_libs
rm -f uri/fetchers/*.lo
rm -f usage/*.o
rm -rf uri/.libs uri/_libs
rm -f usage/*.lo
rm -rf uri/fetchers/.libs uri/fetchers/_libs
rm -f v1/*.o
rm -f v1/*.lo
rm -rf usage/.libs usage/_libs
rm -f version/*.o
rm -rf v1/.libs v1/_libs
rm -f version/*.lo
rm -rf version/.libs version/_libs
rm -f watcher/*.o
rm -rf watcher/.libs watcher/_libs
rm -f watcher/*.lo
rm -rf zookeeper/.libs zookeeper/_libs
rm -f zookeeper/*.o
rm -f zookeeper/*.lo
rm -rf ../include/mesos/.deps ../include/mesos/authentication/.deps 
../include/mesos/authorizer/.deps ../include/mesos/containerizer/.deps 
../include/mesos/executor/.deps ../include/mesos/fetcher/.deps 
../include/mesos/maintenance/.deps ../include/mesos/master/.deps 
../include/mesos/module/.deps ../include/mesos/quota/.deps 
../include/mesos/scheduler/.deps ../include/mesos/slave/.deps 
../include/mesos/uri/.deps ../include/mesos/v1/.deps 
../include/mesos/v1/executor/.deps ../include/mesos/v1/scheduler/.deps 
authentication/cram_md5/.deps authorizer/.deps authorizer/local/.deps cli/.deps 
common/.deps docker/.deps examples/.deps exec/.deps files/.deps hdfs/.deps 
health-check/.deps hook/.deps internal/.deps java/jni/.deps jvm/.deps 
jvm/org/apache/.deps launcher/.deps linux/.deps linux/routing/.deps 
linux/routing/diagnosis/.deps linux/routing/filter/.deps 
linux/routing/link/.deps linux/routing/queueing/.deps local/.deps log/.deps 
log/tool/.deps logging/.deps master/.deps master/allocator/.deps 
master/allocator/mesos/.deps master/allocator/sorter/drf/.deps messages/.deps 
module/.deps sched/.deps scheduler/.deps slave/.deps slave/containerizer/.deps 
slave/containerizer/mesos/.deps 
slave/containerizer/mesos/isolators/cgroups/.deps 
slave/containerizer/mesos/isolators/filesystem/.deps 
slave/containerizer/mesos/isolators/namespaces/.deps 
slave/containerizer/mesos/isolators/network/.deps 
slave/containerizer/mesos/isolators/posix/.deps 
slave/containerizer/mesos/provisioner/.deps 
slave/containerizer/mesos/provisioner/appc/.deps 
slave/containerizer/mesos/provisioner/backends/.deps 
slave/containerizer/mesos/provisioner/docker/.deps slave/qos_controllers/.deps 
slave/resource_estimators/.deps state/.deps tests/.deps tests/common/.deps 
tests/containerizer/.deps uri/.deps uri/fetchers/.deps usage/.deps v1/.deps 
version/.deps watcher/.deps zookeeper/.deps
rm -f Makefile
make[2]: Leaving directory 
`
rm -f config.status config.cache config.log configure.lineno 
config.status.lineno
rm -f Makefile
make[1]: Leaving directory 
`
if test -d "mesos-0.26.0"; then find "mesos-0.26.0" -type d ! -perm -200 -exec 
chmod u+w {} ';' && rm -rf "mesos-0.26.0" || { sleep 5 && rm -rf 
"mesos-0.26.0"; }; else :; fi