Pranith, https://review.gluster.org/c/glusterfs/+/20685 seems to have caused multiple failure runs out of https://review.gluster.org/c/glusterfs/+/20637/8 out of yesterday's report. Did you get a chance to look at it?
On Fri, Aug 10, 2018 at 1:03 PM Pranith Kumar Karampuri <[email protected]> wrote: > > > On Fri, Aug 10, 2018 at 6:34 AM Shyam Ranganathan <[email protected]> > wrote: > >> Today's test results are updated in the spreadsheet in sheet named "Run >> patch set 8". >> >> I took in patch https://review.gluster.org/c/glusterfs/+/20685 which >> caused quite a few failures, so not updating new failures as issue yet. >> >> Please look at the failures for tests that were retried and passed, as >> the logs for the initial runs should be preserved from this run onward. >> >> Otherwise nothing else to report on the run status, if you are averse to >> spreadsheets look at this comment in gerrit [1]. >> >> Shyam >> >> [1] Patch set 8 run status: >> >> https://review.gluster.org/c/glusterfs/+/20637/8#message-54de30fa384fd02b0426d9db6d07fad4eeefcf08 >> On 08/07/2018 07:37 PM, Shyam Ranganathan wrote: >> > Deserves a new beginning, threads on the other mail have gone deep >> enough. >> > >> > NOTE: (5) below needs your attention, rest is just process and data on >> > how to find failures. >> > >> > 1) We are running the tests using the patch [2]. >> > >> > 2) Run details are extracted into a separate sheet in [3] named "Run >> > Failures" use a search to find a failing test and the corresponding run >> > that it failed in. >> > >> > 3) Patches that are fixing issues can be found here [1], if you think >> > you have a patch out there, that is not in this list, shout out. >> > >> > 4) If you own up a test case failure, update the spreadsheet [3] with >> > your name against the test, and also update other details as needed (as >> > comments, as edit rights to the sheet are restricted). >> > >> > 5) Current test failures >> > We still have the following tests failing and some without any RCA or >> > attention, (If something is incorrect, write back). >> > >> > ./tests/bugs/replicate/bug-1290965-detect-bitrotten-objects.t (needs >> > attention) >> > ./tests/00-geo-rep/georep-basic-dr-tarssh.t (Kotresh) >> > ./tests/bugs/glusterd/add-brick-and-validate-replicated-volume-options.t >> > (Atin) >> > ./tests/bugs/ec/bug-1236065.t (Ashish) >> > ./tests/00-geo-rep/georep-basic-dr-rsync.t (Kotresh) >> > ./tests/basic/ec/ec-1468261.t (needs attention) >> > ./tests/basic/afr/add-brick-self-heal.t (needs attention) >> > ./tests/basic/afr/granular-esh/replace-brick.t (needs attention) >> > ./tests/bugs/core/multiplex-limit-issue-151.t (needs attention) >> > ./tests/bugs/glusterd/validating-server-quorum.t (Atin) >> > ./tests/bugs/replicate/bug-1363721.t (Ravi) >> > >> > Here are some newer failures, but mostly one-off failures except cores >> > in ec-5-2.t. All of the following need attention as these are new. >> > >> > ./tests/00-geo-rep/00-georep-verify-setup.t >> > ./tests/basic/afr/gfid-mismatch-resolution-with-fav-child-policy.t >> > ./tests/basic/stats-dump.t >> > ./tests/bugs/bug-1110262.t >> > >> ./tests/bugs/glusterd/mgmt-handshake-and-volume-sync-post-glusterd-restart.t >> > ./tests/basic/ec/ec-data-heal.t >> > ./tests/bugs/replicate/bug-1448804-check-quorum-type-values.t >> > > Sent https://review.gluster.org/c/glusterfs/+/20697 for the test above. > > >> > >> ./tests/bugs/snapshot/bug-1482023-snpashot-issue-with-other-processes-accessing-mounted-path.t >> > ./tests/basic/ec/ec-5-2.t >> > >> > 6) Tests that are addressed or are not occurring anymore are, >> > >> > ./tests/bugs/glusterd/rebalance-operations-in-single-node.t >> > ./tests/bugs/index/bug-1559004-EMLINK-handling.t >> > ./tests/bugs/replicate/bug-1386188-sbrain-fav-child.t >> > ./tests/bugs/replicate/bug-1433571-undo-pending-only-on-up-bricks.t >> > ./tests/bitrot/bug-1373520.t >> > ./tests/bugs/distribute/bug-1117851.t >> > ./tests/bugs/glusterd/quorum-validation.t >> > ./tests/bugs/distribute/bug-1042725.t >> > >> ./tests/bugs/replicate/bug-1586020-mark-dirty-for-entry-txn-on-quorum-failure.t >> > ./tests/bugs/quota/bug-1293601.t >> > ./tests/bugs/bug-1368312.t >> > ./tests/bugs/distribute/bug-1122443.t >> > ./tests/bugs/core/bug-1432542-mpx-restart-crash.t >> > >> > Shyam (and Atin) >> > >> > On 08/05/2018 06:24 PM, Shyam Ranganathan wrote: >> >> Health on master as of the last nightly run [4] is still the same. >> >> >> >> Potential patches that rectify the situation (as in [1]) are bunched in >> >> a patch [2] that Atin and myself have put through several regressions >> >> (mux, normal and line coverage) and these have also not passed. >> >> >> >> Till we rectify the situation we are locking down master branch commit >> >> rights to the following people, Amar, Atin, Shyam, Vijay. >> >> >> >> The intention is to stabilize master and not add more patches that my >> >> destabilize it. >> >> >> >> Test cases that are tracked as failures and need action are present >> here >> >> [3]. >> >> >> >> @Nigel, request you to apply the commit rights change as you see this >> >> mail and let the list know regarding the same as well. >> >> >> >> Thanks, >> >> Shyam >> >> >> >> [1] Patches that address regression failures: >> >> https://review.gluster.org/#/q/starredby:srangana%2540redhat.com >> >> >> >> [2] Bunched up patch against which regressions were run: >> >> https://review.gluster.org/#/c/20637 >> >> >> >> [3] Failing tests list: >> >> >> https://docs.google.com/spreadsheets/d/1IF9GhpKah4bto19RQLr0y_Kkw26E_-crKALHSaSjZMQ/edit?usp=sharing >> >> >> >> [4] Nightly run dashboard: >> https://build.gluster.org/job/nightly-master/ >> > _______________________________________________ >> > Gluster-devel mailing list >> > [email protected] >> > https://lists.gluster.org/mailman/listinfo/gluster-devel >> > >> _______________________________________________ >> maintainers mailing list >> [email protected] >> https://lists.gluster.org/mailman/listinfo/maintainers >> > > > -- > Pranith > _______________________________________________ > maintainers mailing list > [email protected] > https://lists.gluster.org/mailman/listinfo/maintainers >
_______________________________________________ Gluster-devel mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-devel
