Paul Burba <ptbu...@gmail.com> writes: > Hi All, > > One of the roadmap items yet to be started was a 'test review' to > 'Determine which XFail and WIP tests should remain so, and which need > to be fixed before release.' > > I took a look at this today. We currently have 61 tests set to XFail > (2 of these are WIPs). Here is how they break down: > > A) Tests with an associated issue (35). > > The target milestone of these 35 break down like this: > > (8) Never scheduled, i.e. '---' > (9) Unscheduled [3 of these are marked as RESOLVED] > (1) 1.6.1 > (3) 1.7.0 > (8) 1.7-consider > (1) 1.8.0 > (5) 1.8.0-consider > > B) 26 tests have no issue associated with them (well there *might* be > an issue, but a cursory look in the issue tracker didn't reveal > anything obvious). > > So a whopping 55 tests are *potentially* release blockers (all but the > 1.8.0* milestone issues). > > I put a summary of all 61 Xfailing tests in > ^/subversion/trunk/notes/xfail-status[1]. > > For many of the tests I listed a "point person" as an educated guess > as to who is best equipped to decide the fate of the test for 1.7. In > the case of tests with an associated issue that is assigned to > someone, the point person is the assignee. For issues not assigned to > anyone, I left this blank. In the case of tests with no associated > issues, this is usually the person who committed the test or changed > it to XFail. If I cc'ed you then you are one of the lucky point > people :-P > > Keep in mind this is just a guess, if your name is there you aren't > compelled to do anything...though I suspect we'd all agree it's bad > form to create an XFailing test with no associated issue when we > aren't planning on fixing the underlying problem. > > If you have some time please look through the list. If a test you > wrote needs an issue please write one up. If there is an existing > issue that should be associated with a test add it. If an associated > issue is unscheduled or never scheduled and you're familiar with it, > please take a stab at some issue triage. > > Thanks, > > Paul > > [1] I'm happy to find this a new home if it belongs elsewhere. > Ideally I'd like to tweak svntest\testcase.py:XFail() so that we have > a way to associate an issue and a target milestone with each XFailing > test, so we can easily create this list. I'll take a look at doing > just that tomorrow once I get through the items I've assigned myself.
LISTING: depth_tests.py Test # Mode Test Description ------ ----- ---------------- 36 XFAIL 'info' should treat excluded item as versioned Issue #: None Target Milestone: N/A Thread: N/A Log: r876772 Point Person: danielsh I think issue #3787 is related to this XFail test. Thanks and Regards Noorul