See <https://build.gluster.org/job/regression-test-with-multiplex/1647/display/redirect>
Changes: ------------------------------------------ [...truncated 2.43 MB...] Files=1, Tests=15, 24 wallclock secs ( 0.02 usr 0.01 sys + 1.05 cusr 0.70 csys = 1.78 CPU) Result: PASS Logs preserved in tarball rebalance-in-cluster-iteration-1.tar End of test ./tests/bugs/glusterd/rebalance-in-cluster.t ================================================================================ ================================================================================ [22:01:26] Running tests in file ./tests/bugs/glusterd/rebalance-operations-in-single-node.t ./tests/bugs/glusterd/rebalance-operations-in-single-node.t .. 1..48 ok 1 [ 281/ 1377] < 16> 'glusterd' ok 2 [ 11/ 21] < 17> 'pidof glusterd' No volumes present ok 3 [ 11/ 197] < 19> 'gluster --mode=script --wignore volume info' ok 4 [ 24/ 116] < 20> 'gluster --mode=script --wignore volume create StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest1 builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest2 builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest3 builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest4' ok 5 [ 29/ 1517] < 21> 'gluster --mode=script --wignore volume start StartMigrationDuringRebalanceTest' ok 6 [ 80/ 5884] < 24> 'gluster --mode=script --wignore volume rebalance StartMigrationDuringRebalanceTest start' ok 7 [ 90/ 105] < 29> '! gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest1 status' ok 8 [ 17/ 85] < 31> 'completed rebalance_status_field StartMigrationDuringRebalanceTest' ok 9 [ 44/ 5225] < 33> 'gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest1 start' ok 10 [ 11/ 144] < 34> '! gluster --mode=script --wignore volume rebalance StartMigrationDuringRebalanceTest start' ok 11 [ 40/ 71] < 35> '! gluster --mode=script --wignore volume rebalance StartMigrationDuringRebalanceTest status' ok 12 [ 12/ 69] < 36> '! gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest2 start' ok 13 [ 11/ 103] < 38> 'completed remove_brick_status_completed_field StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest1' ok 14 [ 11/ 180] < 40> 'gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest1 commit' ok 15 [ 14/ 5089] < 42> 'gluster --mode=script --wignore volume rebalance StartMigrationDuringRebalanceTest start' ok 16 [ 32/ 63] < 43> 'completed rebalance_status_field StartMigrationDuringRebalanceTest' ok 17 [ 9/ 49] < 44> 'gluster --mode=script --wignore volume rebalance StartMigrationDuringRebalanceTest stop' ok 18 [ 9/ 5080] < 46> 'gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest2 start' ok 19 [ 37/ 78] < 47> 'gluster --mode=script --wignore volume remove-brick StartMigrationDuringRebalanceTest builder210.int.aws.gluster.org:/d/backends/StartMigrationDuringRebalanceTest2 stop' ok 20 [ 9/ 81] < 51> 'gluster --mode=script --wignore volume create patchy builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3' ok 21 [ 14/ 1194] < 52> 'gluster --mode=script --wignore volume start patchy' ok 22 [ 12/ 55] < 55> 'glusterfs -s builder210.int.aws.gluster.org --volfile-id patchy /mnt/glusterfs/0' ok 23 [ 12/ 62] < 56> 'mkdir /mnt/glusterfs/0/dir1 /mnt/glusterfs/0/dir2 /mnt/glusterfs/0/dir3 /mnt/glusterfs/0/dir4 /mnt/glusterfs/0/dir5 /mnt/glusterfs/0/dir6 /mnt/glusterfs/0/dir7 /mnt/glusterfs/0/dir8 /mnt/glusterfs/0/dir9 /mnt/glusterfs/0/dir10' ok 24 [ 14/ 157] < 57> 'touch /mnt/glusterfs/0/dir1/file1 /mnt/glusterfs/0/dir1/file2 /mnt/glusterfs/0/dir1/file3 /mnt/glusterfs/0/dir1/file4 /mnt/glusterfs/0/dir1/file5 /mnt/glusterfs/0/dir1/file6 /mnt/glusterfs/0/dir1/file7 /mnt/glusterfs/0/dir1/file8 /mnt/glusterfs/0/dir1/file9 /mnt/glusterfs/0/dir1/file10 /mnt/glusterfs/0/dir2/file1 /mnt/glusterfs/0/dir2/file2 /mnt/glusterfs/0/dir2/file3 /mnt/glusterfs/0/dir2/file4 /mnt/glusterfs/0/dir2/file5 /mnt/glusterfs/0/dir2/file6 /mnt/glusterfs/0/dir2/file7 /mnt/glusterfs/0/dir2/file8 /mnt/glusterfs/0/dir2/file9 /mnt/glusterfs/0/dir2/file10 /mnt/glusterfs/0/dir3/file1 /mnt/glusterfs/0/dir3/file2 /mnt/glusterfs/0/dir3/file3 /mnt/glusterfs/0/dir3/file4 /mnt/glusterfs/0/dir3/file5 /mnt/glusterfs/0/dir3/file6 /mnt/glusterfs/0/dir3/file7 /mnt/glusterfs/0/dir3/file8 /mnt/glusterfs/0/dir3/file9 /mnt/glusterfs/0/dir3/file10 /mnt/glusterfs/0/dir4/file1 /mnt/glusterfs/0/dir4/file2 /mnt/glusterfs/0/dir4/file3 /mnt/glusterfs/0/dir4/fi le4 /mnt/glusterfs/0/dir4/file5 /mnt/glusterfs/0/dir4/file6 /mnt/glusterfs/0/dir4/file7 /mnt/glusterfs/0/dir4/file8 /mnt/glusterfs/0/dir4/file9 /mnt/glusterfs/0/dir4/file10 /mnt/glusterfs/0/dir5/file1 /mnt/glusterfs/0/dir5/file2 /mnt/glusterfs/0/dir5/file3 /mnt/glusterfs/0/dir5/file4 /mnt/glusterfs/0/dir5/file5 /mnt/glusterfs/0/dir5/file6 /mnt/glusterfs/0/dir5/file7 /mnt/glusterfs/0/dir5/file8 /mnt/glusterfs/0/dir5/file9 /mnt/glusterfs/0/dir5/file10 /mnt/glusterfs/0/dir6/file1 /mnt/glusterfs/0/dir6/file2 /mnt/glusterfs/0/dir6/file3 /mnt/glusterfs/0/dir6/file4 /mnt/glusterfs/0/dir6/file5 /mnt/glusterfs/0/dir6/file6 /mnt/glusterfs/0/dir6/file7 /mnt/glusterfs/0/dir6/file8 /mnt/glusterfs/0/dir6/file9 /mnt/glusterfs/0/dir6/file10 /mnt/glusterfs/0/dir7/file1 /mnt/glusterfs/0/dir7/file2 /mnt/glusterfs/0/dir7/file3 /mnt/glusterfs/0/dir7/file4 /mnt/glusterfs/0/dir7/file5 /mnt/glusterfs/0/dir7/file6 /mnt/glusterfs/0/dir7/file7 /mnt/glusterfs/0/dir7/file8 /mnt/glusterfs/0/dir7/file9 /m nt/glusterfs/0/dir7/file10 /mnt/glusterfs/0/dir8/file1 /mnt/glusterfs/0/dir8/file2 /mnt/glusterfs/0/dir8/file3 /mnt/glusterfs/0/dir8/file4 /mnt/glusterfs/0/dir8/file5 /mnt/glusterfs/0/dir8/file6 /mnt/glusterfs/0/dir8/file7 /mnt/glusterfs/0/dir8/file8 /mnt/glusterfs/0/dir8/file9 /mnt/glusterfs/0/dir8/file10 /mnt/glusterfs/0/dir9/file1 /mnt/glusterfs/0/dir9/file2 /mnt/glusterfs/0/dir9/file3 /mnt/glusterfs/0/dir9/file4 /mnt/glusterfs/0/dir9/file5 /mnt/glusterfs/0/dir9/file6 /mnt/glusterfs/0/dir9/file7 /mnt/glusterfs/0/dir9/file8 /mnt/glusterfs/0/dir9/file9 /mnt/glusterfs/0/dir9/file10 /mnt/glusterfs/0/dir10/file1 /mnt/glusterfs/0/dir10/file2 /mnt/glusterfs/0/dir10/file3 /mnt/glusterfs/0/dir10/file4 /mnt/glusterfs/0/dir10/file5 /mnt/glusterfs/0/dir10/file6 /mnt/glusterfs/0/dir10/file7 /mnt/glusterfs/0/dir10/file8 /mnt/glusterfs/0/dir10/file9 /mnt/glusterfs/0/dir10/file10' ok 25 [ 11/ 274] < 60> 'gluster --mode=script --wignore volume add-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy4' ok 26 [ 27/ 5113] < 61> 'gluster --mode=script --wignore volume rebalance patchy start' ok 27 [ 83/ 2092] < 62> 'completed rebalance_status_field patchy' ok 28 [ 535/ 2245] < 74> 'glusterd' ok 29 [ 424/ 1] < 83> '[ 18 == 18 ]' ok 30 [ 12/ 1] < 84> '[ 0Bytes == 0Bytes ]' ok 31 [ 12/ 1] < 85> '[ 100 == 100 ]' ok 32 [ 13/ 1] < 86> '[ 0 == 0 ]' ok 33 [ 12/ 1] < 87> '[ 0 == 0 ]' ok 34 [ 14/ 24] < 89> 'Y force_umount /mnt/glusterfs/0' ok 35 [ 12/ 1144] < 93> 'gluster --mode=script --wignore volume start patchy force' ok 36 [ 17/ 88] < 94> 'glusterfs -s builder210.int.aws.gluster.org --volfile-id patchy /mnt/glusterfs/0' ok 37 [ 10996/ 424] < 108> 'gluster --mode=script --wignore volume add-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy5 builder210.int.aws.gluster.org:/d/backends/patchy6' ok 38 [ 24/ 5266] < 111> 'gluster --mode=script --wignore volume rebalance patchy fix-layout start' ok 39 [ 28/ 26682] < 113> 'fix-layout completed fix-layout_status_field patchy' ok 40 [ 15/ 111] < 116> 'mkdir /mnt/glusterfs/0/dir21 /mnt/glusterfs/0/dir22 /mnt/glusterfs/0/dir23 /mnt/glusterfs/0/dir24 /mnt/glusterfs/0/dir25 /mnt/glusterfs/0/dir26 /mnt/glusterfs/0/dir27 /mnt/glusterfs/0/dir28 /mnt/glusterfs/0/dir29 /mnt/glusterfs/0/dir30' ok 41 [ 44/ 197] < 117> 'touch /mnt/glusterfs/0/dir21/files1 /mnt/glusterfs/0/dir21/files2 /mnt/glusterfs/0/dir21/files3 /mnt/glusterfs/0/dir21/files4 /mnt/glusterfs/0/dir21/files5 /mnt/glusterfs/0/dir21/files6 /mnt/glusterfs/0/dir21/files7 /mnt/glusterfs/0/dir21/files8 /mnt/glusterfs/0/dir21/files9 /mnt/glusterfs/0/dir21/files10 /mnt/glusterfs/0/dir22/files1 /mnt/glusterfs/0/dir22/files2 /mnt/glusterfs/0/dir22/files3 /mnt/glusterfs/0/dir22/files4 /mnt/glusterfs/0/dir22/files5 /mnt/glusterfs/0/dir22/files6 /mnt/glusterfs/0/dir22/files7 /mnt/glusterfs/0/dir22/files8 /mnt/glusterfs/0/dir22/files9 /mnt/glusterfs/0/dir22/files10 /mnt/glusterfs/0/dir23/files1 /mnt/glusterfs/0/dir23/files2 /mnt/glusterfs/0/dir23/files3 /mnt/glusterfs/0/dir23/files4 /mnt/glusterfs/0/dir23/files5 /mnt/glusterfs/0/dir23/files6 /mnt/glusterfs/0/dir23/files7 /mnt/glusterfs/0/dir23/files8 /mnt/glusterfs/0/dir23/files9 /mnt/glusterfs/0/dir23/files10 /mnt/glusterfs/0/dir24/files1 /mnt/glusterfs/0/d ir24/files2 /mnt/glusterfs/0/dir24/files3 /mnt/glusterfs/0/dir24/files4 /mnt/glusterfs/0/dir24/files5 /mnt/glusterfs/0/dir24/files6 /mnt/glusterfs/0/dir24/files7 /mnt/glusterfs/0/dir24/files8 /mnt/glusterfs/0/dir24/files9 /mnt/glusterfs/0/dir24/files10 /mnt/glusterfs/0/dir25/files1 /mnt/glusterfs/0/dir25/files2 /mnt/glusterfs/0/dir25/files3 /mnt/glusterfs/0/dir25/files4 /mnt/glusterfs/0/dir25/files5 /mnt/glusterfs/0/dir25/files6 /mnt/glusterfs/0/dir25/files7 /mnt/glusterfs/0/dir25/files8 /mnt/glusterfs/0/dir25/files9 /mnt/glusterfs/0/dir25/files10 /mnt/glusterfs/0/dir26/files1 /mnt/glusterfs/0/dir26/files2 /mnt/glusterfs/0/dir26/files3 /mnt/glusterfs/0/dir26/files4 /mnt/glusterfs/0/dir26/files5 /mnt/glusterfs/0/dir26/files6 /mnt/glusterfs/0/dir26/files7 /mnt/glusterfs/0/dir26/files8 /mnt/glusterfs/0/dir26/files9 /mnt/glusterfs/0/dir26/files10 /mnt/glusterfs/0/dir27/files1 /mnt/glusterfs/0/dir27/files2 /mnt/glusterfs/0/dir27/files3 /mnt/glusterfs/0/dir27/files4 /mnt/glusterfs /0/dir27/files5 /mnt/glusterfs/0/dir27/files6 /mnt/glusterfs/0/dir27/files7 /mnt/glusterfs/0/dir27/files8 /mnt/glusterfs/0/dir27/files9 /mnt/glusterfs/0/dir27/files10 /mnt/glusterfs/0/dir28/files1 /mnt/glusterfs/0/dir28/files2 /mnt/glusterfs/0/dir28/files3 /mnt/glusterfs/0/dir28/files4 /mnt/glusterfs/0/dir28/files5 /mnt/glusterfs/0/dir28/files6 /mnt/glusterfs/0/dir28/files7 /mnt/glusterfs/0/dir28/files8 /mnt/glusterfs/0/dir28/files9 /mnt/glusterfs/0/dir28/files10 /mnt/glusterfs/0/dir29/files1 /mnt/glusterfs/0/dir29/files2 /mnt/glusterfs/0/dir29/files3 /mnt/glusterfs/0/dir29/files4 /mnt/glusterfs/0/dir29/files5 /mnt/glusterfs/0/dir29/files6 /mnt/glusterfs/0/dir29/files7 /mnt/glusterfs/0/dir29/files8 /mnt/glusterfs/0/dir29/files9 /mnt/glusterfs/0/dir29/files10 /mnt/glusterfs/0/dir30/files1 /mnt/glusterfs/0/dir30/files2 /mnt/glusterfs/0/dir30/files3 /mnt/glusterfs/0/dir30/files4 /mnt/glusterfs/0/dir30/files5 /mnt/glusterfs/0/dir30/files6 /mnt/glusterfs/0/dir30/files7 /mnt/glust erfs/0/dir30/files8 /mnt/glusterfs/0/dir30/files9 /mnt/glusterfs/0/dir30/files10' ok 42 [ 13/ 468] < 119> 'gluster --mode=script --wignore volume add-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy7 builder210.int.aws.gluster.org:/d/backends/patchy8' ok 43 [ 201/ 5374] < 121> 'gluster --mode=script --wignore volume rebalance patchy start force' ok 44 [ 172/ 64556] < 122> 'completed rebalance_status_field patchy' ok 45 [ 13/ 48] < 124> 'pkill gluster' ok 46 [ 23/ 1598] < 125> 'glusterd' ok 47 [ 32/ 41] < 126> 'pidof glusterd' ok 48 [ 19/ 85] < 129> 'completed rebalance_status_field patchy' ok All tests successful. Files=1, Tests=48, 158 wallclock secs ( 0.03 usr 0.00 sys + 12.24 cusr 8.05 csys = 20.32 CPU) Result: PASS Logs preserved in tarball rebalance-operations-in-single-node-iteration-1.tar End of test ./tests/bugs/glusterd/rebalance-operations-in-single-node.t ================================================================================ ================================================================================ [22:04:04] Running tests in file ./tests/bugs/glusterd/remove-brick-in-cluster.t ./tests/bugs/glusterd/remove-brick-in-cluster.t .. 1..21 ok 1 [ 236/ 2418] < 8> 'launch_cluster 2' ok 2 [ 14/ 173] < 11> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume create patchy replica 2 127.1.1.1:/d/backends/1/patchy1 127.1.1.1:/d/backends/1/patchy2 127.1.1.1:/d/backends/1/patchy3 127.1.1.1:/d/backends/1/patchy4' ok 3 [ 12/ 2279] < 12> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume start patchy' ok 4 [ 31/ 156] < 14> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log peer probe 127.1.1.2' ok 5 [ 17/ 1343] < 15> '1 peer_count' ok 6 [ 11/ 5179] < 17> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/2/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli2.log volume remove-brick patchy 127.1.1.1:/d/backends/1/patchy3 127.1.1.1:/d/backends/1/patchy4 start' ok 7 [ 28/ 94] < 18> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/2/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli2.log volume info' ok 8 [ 11/ 143] < 21> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume create patchy1 127.1.1.1:/d/backends/1/patchy10 127.1.1.2:/d/backends/2/patchy11' ok 9 [ 16/ 306] < 22> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume start patchy1' ok 10 [ 11/ 5152] < 23> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume remove-brick patchy1 127.1.1.2:/d/backends/2/patchy11 start' ok 11 [ 29/ 72] < 24> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume status' ok 12 [ 11/ 3128] < 26> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume stop patchy' ok 13 [ 10/ 64] < 27> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume delete patchy' ok 14 [ 10/ 205] < 31> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume create patchy replica 3 127.1.1.1:/d/backends/1/brick1 127.1.1.2:/d/backends/2/brick2 127.1.1.1:/d/backends/1/brick3 127.1.1.2:/d/backends/2/brick4 127.1.1.1:/d/backends/1/brick5 127.1.1.2:/d/backends/2/brick6' ok 15 [ 11/ 3283] < 36> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume start patchy' ok 16 [ 29/ 269] < 39> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume remove-brick patchy replica 2 127.1.1.1:/d/backends/1/brick1 127.1.1.2:/d/backends/2/brick6 force' ok 17 [ 47/ 227] < 42> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume remove-brick patchy replica 2 127.1.1.1:/d/backends/1/brick3 127.1.1.2:/d/backends/2/brick2 force' ok 18 [ 15/ 3115] < 45> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume remove-brick patchy replica 1 127.1.1.1:/d/backends/1/brick5 force' ok 19 [ 12/ 2286] < 51> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume add-brick patchy replica 2 127.1.1.1:/d/backends/1/brick5 force' ok 20 [ 12/ 260] < 54> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume add-brick patchy replica 2 127.1.1.1:/d/backends/1/brick3 127.1.1.2:/d/backends/2/brick2 force' ok 21 [ 14/ 532] < 57> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/remove-brick-in-cluster.t_cli1.log volume add-brick patchy replica 3 127.1.1.1:/d/backends/1/brick1 127.1.1.2:/d/backends/2/brick6 force' ok All tests successful. Files=1, Tests=21, 32 wallclock secs ( 0.02 usr 0.01 sys + 1.34 cusr 0.84 csys = 2.21 CPU) Result: PASS Logs preserved in tarball remove-brick-in-cluster-iteration-1.tar End of test ./tests/bugs/glusterd/remove-brick-in-cluster.t ================================================================================ ================================================================================ [22:04:36] Running tests in file ./tests/bugs/glusterd/remove-brick-testcases.t ./tests/bugs/glusterd/remove-brick-testcases.t .. 1..35 ok 1 [ 216/ 1313] < 20> 'glusterd' ok 2 [ 12/ 9] < 21> 'pidof glusterd' ok 3 [ 10/ 258] < 23> 'gluster --mode=script --wignore volume create patchy builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 builder210.int.aws.gluster.org:/d/backends/patchy4 builder210.int.aws.gluster.org:/d/backends/patchy5' ok 4 [ 51/ 1509] < 24> 'gluster --mode=script --wignore volume start patchy' OK ok 5 [ 2267/ 207] < 29> '0 brick_up_status patchy builder210.int.aws.gluster.org /d/backends/patchy1' ok 6 [ 78/ 137] < 32> '! gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy1 start' ok 7 [ 84/ 520] < 34> 'gluster --mode=script --wignore volume start patchy force' ok 8 [ 78/ 404] < 35> '1 brick_up_status patchy builder210.int.aws.gluster.org /d/backends/patchy1' ok 9 [ 283/ 5527] < 38> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy1 start' ok 10 [ 183/ 830] < 40> 'completed remove_brick_status_completed_field patchy builder210.int.aws.gluster.org:/d/backends/patchy1' OK ok 11 [ 2392/ 397] < 46> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy1 commit' ok 12 [ 72/ 5629] < 50> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 start' ok 13 [ 80/ 127] < 54> '! gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/ABCD status' ok 14 [ 12/ 165] < 55> '! gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/ABCD stop' ok 15 [ 12/ 81] < 60> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 status' ok 16 [ 83/ 258] < 61> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 stop' ok 17 [ 11/ 181] < 64> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 force' ok 18 [ 39/ 136] < 65> '3 brick_count patchy' ok 19 [ 11/ 163] < 67> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy3 force' ok 20 [ 12/ 144] < 68> '2 brick_count patchy' ok 21 [ 29/ 87] < 76> 'decommissioned remove_brick_commit_status' ok 22 [ 81/ 1124] < 78> 'gluster --mode=script --wignore volume stop patchy' ok 23 [ 20/ 3961] < 79> 'gluster --mode=script --wignore volume delete patchy' ok 24 [ 18/ 182] < 82> 'gluster --mode=script --wignore volume create patchy replica 3 builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 builder210.int.aws.gluster.org:/d/backends/patchy4 builder210.int.aws.gluster.org:/d/backends/patchy5 builder210.int.aws.gluster.org:/d/backends/patchy6' ok 25 [ 35/ 2505] < 83> 'gluster --mode=script --wignore volume start patchy' ok 26 [ 85/ 259] < 90> 'failed remove_brick_start_status' ok 27 [ 270/ 323] < 98> 'decommissioned remove_brick_commit_status2' ok 28 [ 374/ 251] < 99> 'gluster --mode=script --wignore volume info patchy' ok 29 [ 88/ 394] < 108> 'success remove_brick_status' ok 30 [ 206/ 253] < 109> 'gluster --mode=script --wignore volume info patchy' ok 31 [ 133/ 5482] < 113> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy4 builder210.int.aws.gluster.org:/d/backends/patchy5 start' ok 32 [ 192/ 1512] < 114> 'completed remove_brick_status_completed_field patchy builder210.int.aws.gluster.org:/d/backends/patchy5' ok 33 [ 10/ 76] < 115> 'completed remove_brick_status_completed_field patchy builder210.int.aws.gluster.org:/d/backends/patchy4' ok 34 [ 25/ 329] < 116> 'gluster --mode=script --wignore volume remove-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy4 builder210.int.aws.gluster.org:/d/backends/patchy5 commit' ok 35 [ 110/ 1420] < 117> 'gluster --mode=script --wignore volume remove-brick patchy replica 1 builder210.int.aws.gluster.org:/d/backends/patchy2 force' ok All tests successful. Files=1, Tests=35, 46 wallclock secs ( 0.03 usr 0.00 sys + 2.25 cusr 1.35 csys = 3.63 CPU) Result: PASS Logs preserved in tarball remove-brick-testcases-iteration-1.tar End of test ./tests/bugs/glusterd/remove-brick-testcases.t ================================================================================ ================================================================================ [22:05:22] Running tests in file ./tests/bugs/glusterd/removing-multiple-bricks-in-single-remove-brick-command.t ./tests/bugs/glusterd/removing-multiple-bricks-in-single-remove-brick-command.t .. 1..21 ok 1 [ 1316/ 1963] < 9> 'glusterd' ok 2 [ 74/ 59] < 10> 'pidof glusterd' No volumes present ok 3 [ 222/ 450] < 11> 'gluster --mode=script --wignore volume info' ok 4 [ 120/ 847] < 14> 'gluster --mode=script --wignore volume create patchy replica 2 builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 builder210.int.aws.gluster.org:/d/backends/patchy4 builder210.int.aws.gluster.org:/d/backends/patchy5 builder210.int.aws.gluster.org:/d/backends/patchy6' ok 5 [ 84/ 2139] < 15> 'gluster --mode=script --wignore volume start patchy' ok 6 [ 132/ 319] < 19> 'glusterfs -s builder210.int.aws.gluster.org --volfile-id patchy /mnt/glusterfs/0' ok 7 [ 115/ 793] < 20> 'touch /mnt/glusterfs/0/file1 /mnt/glusterfs/0/file2 /mnt/glusterfs/0/file3 /mnt/glusterfs/0/file4 /mnt/glusterfs/0/file5 /mnt/glusterfs/0/file6 /mnt/glusterfs/0/file7 /mnt/glusterfs/0/file8 /mnt/glusterfs/0/file9 /mnt/glusterfs/0/file10' ok 8 [ 21/ 5777] < 29> 'success remove_brick_start_status' ok 9 [ 342/ 2183] < 32> 'completed remove_brick_status_completed_field patchy builder210.int.aws.gluster.org:/d/backends/patchy6 builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy5' ok 10 [ 95/ 467] < 40> 'success remove_brick_commit_status' ok 11 [ 660/ 101] < 44> 'Replicate echo Replicate' ok 12 [ 501/ 453] < 45> 'Y force_umount /mnt/glusterfs/0' ok 13 [ 144/ 1798] < 50> 'gluster --mode=script --wignore volume create patchy1 replica 3 builder210.int.aws.gluster.org:/d/backends/patchy10 builder210.int.aws.gluster.org:/d/backends/patchy11 builder210.int.aws.gluster.org:/d/backends/patchy12 builder210.int.aws.gluster.org:/d/backends/patchy13 builder210.int.aws.gluster.org:/d/backends/patchy14 builder210.int.aws.gluster.org:/d/backends/patchy15 builder210.int.aws.gluster.org:/d/backends/patchy16 builder210.int.aws.gluster.org:/d/backends/patchy17 builder210.int.aws.gluster.org:/d/backends/patchy18' ok 14 [ 71/ 2784] < 51> 'gluster --mode=script --wignore volume start patchy1' ok 15 [ 71/ 577] < 54> 'glusterfs -s builder210.int.aws.gluster.org --volfile-id patchy1 /mnt/glusterfs/0' ok 16 [ 95/ 35] < 55> 'touch /mnt/glusterfs/0/zerobytefile.txt' ok 17 [ 104/ 119] < 56> 'mkdir /mnt/glusterfs/0/test_dir' ok 18 [ 77/ 240] < 57> 'dd if=/dev/zero of=/mnt/glusterfs/0/file bs=1024 count=1024' ok 19 [ 76/ 276] < 70> 'failed remove_brick_start' ok 20 [ 746/ 2239] < 75> 'success remove_brick' ok 21 [ 171/ 222] < 77> 'Y force_umount /mnt/glusterfs/0' ok All tests successful. Files=1, Tests=21, 31 wallclock secs ( 0.03 usr 0.00 sys + 1.21 cusr 1.04 csys = 2.28 CPU) Result: PASS Logs preserved in tarball removing-multiple-bricks-in-single-remove-brick-command-iteration-1.tar End of test ./tests/bugs/glusterd/removing-multiple-bricks-in-single-remove-brick-command.t ================================================================================ ================================================================================ [22:05:54] Running tests in file ./tests/bugs/glusterd/replace-brick-operations.t ./tests/bugs/glusterd/replace-brick-operations.t .. 1..14 ok 1 [ 986/ 1403] < 11> 'glusterd' ok 2 [ 11/ 17] < 12> 'pidof glusterd' ok 3 [ 11/ 272] < 15> 'gluster --mode=script --wignore volume create patchy replica 2 builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy2' ok 4 [ 89/ 2738] < 16> 'gluster --mode=script --wignore volume start patchy' ok 5 [ 166/ 286] < 24> '! gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 start' ok 6 [ 122/ 252] < 25> '! gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 status' ok 7 [ 77/ 175] < 26> '! gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 abort' ok 8 [ 76/ 3838] < 30> 'gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy2 builder210.int.aws.gluster.org:/d/backends/patchy3 commit force' ok 9 [ 30/ 114] < 34> 'glusterfs --volfile-id=patchy --volfile-server=builder210.int.aws.gluster.org /mnt/glusterfs/0' ok 10 [ 25/ 2262] < 37> 'gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy1 builder210.int.aws.gluster.org:/d/backends/patchy1_new commit force' ok 11 [ 68/ 237] < 39> '1 afr_child_up_status patchy 1' ok 12 [ 12/ 2101] < 41> 'kill_brick patchy builder210.int.aws.gluster.org /d/backends/patchy1_new' ok 13 [ 12/ 2222] < 44> 'gluster --mode=script --wignore volume replace-brick patchy builder210.int.aws.gluster.org:/d/backends/patchy1_new builder210.int.aws.gluster.org:/d/backends/patchy1_newer commit force' ok 14 [ 11/ 442] < 46> '1 afr_child_up_status patchy 1' ok All tests successful. Files=1, Tests=14, 18 wallclock secs ( 0.02 usr 0.00 sys + 0.96 cusr 0.72 csys = 1.70 CPU) Result: PASS Logs preserved in tarball replace-brick-operations-iteration-1.tar End of test ./tests/bugs/glusterd/replace-brick-operations.t ================================================================================ ================================================================================ [22:06:12] Running tests in file ./tests/bugs/glusterd/reset-brick-and-daemons-follow-quorum.t ./tests/bugs/glusterd/reset-brick-and-daemons-follow-quorum.t .. 1..17 ok 1 [ 224/ 4077] < 24> 'launch_cluster 3' ok 2 [ 11/ 132] < 25> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log peer probe 127.1.1.2' ok 3 [ 18/ 88] < 26> '1 check_peers' ok 4 [ 12/ 138] < 28> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log volume create patchy replica 2 127.1.1.1:/d/backends/patchy 127.1.1.2:/d/backends/patchy' ok 5 [ 12/ 2506] < 29> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log volume start patchy' ok 6 [ 17/ 83] < 33> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log volume reset-brick patchy 127.1.1.1:/d/backends/patchy 127.1.1.1:/d/backends/patchy commit force' ok 7 [ 12/ 1070] < 35> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log volume reset-brick patchy 127.1.1.1:/d/backends/patchy start' ok 8 [ 11/ 4238] < 37> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log volume reset-brick patchy 127.1.1.1:/d/backends/patchy 127.1.1.1:/d/backends/patchy commit force' ok 9 [ 56/ 120] < 41> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/reset-brick-and-daemons-follow-quorum.t_cli1.log peer probe 127.1.1.3' ok 10 [ 12/ 1373] < 42> '2 peer_count' ok 11 [ 13/ 78] < 44> '1 cluster_brick_up_status 1 patchy 127.1.1.1 /d/backends/patchy' ok 12 [ 11/ 80] < 45> '1 cluster_brick_up_status 1 patchy 127.1.1.2 /d/backends/patchy' ok 13 [ 11/ 78] < 46> 'Y shd_up_status_1' ok 14 [ 11/ 82] < 47> 'Y shd_up_status_2' ok 15 [ 96/ 3] < 53> 'kill_glusterd 1' ok 16 [ 12/ 1510] < 56> 'glusterd --xlator-option management.working-directory=/d/backends/1/glusterd --xlator-option management.transport.socket.bind-address=127.1.1.1 --xlator-option management.run-directory=/d/backends/1/run/gluster --xlator-option management.glusterd-sockfile=/d/backends/1/glusterd/gd.sock --xlator-option management.cluster-test-mode=/var/log/glusterfs/1 --log-file=/var/log/glusterfs/1/reset-brick-and-daemons-follow-quorum.t_glusterd1.log --pid-file=/d/backends/1/glusterd.pid' ok 17 [ 28/ 1758] < 61> 'Y shd_up_status_2' ok All tests successful. Files=1, Tests=17, 18 wallclock secs ( 0.02 usr 0.00 sys + 1.38 cusr 0.90 csys = 2.30 CPU) Result: PASS Logs preserved in tarball reset-brick-and-daemons-follow-quorum-iteration-1.tar End of test ./tests/bugs/glusterd/reset-brick-and-daemons-follow-quorum.t ================================================================================ ================================================================================ [22:06:31] Running tests in file ./tests/bugs/glusterd/serialize-shd-manager-glusterd-restart.t Process leaked file descriptors. See https://jenkins.io/redirect/troubleshooting/process-leaked-file-descriptors for more information Build step 'Execute shell' marked build as failure _______________________________________________ maintainers mailing list maintainers@gluster.org https://lists.gluster.org/mailman/listinfo/maintainers