Hello XWiki Community, I would like to propose a change in the visibility of manual testing.
The Problem: As you know, we have a manual test report found here: http://www.xwiki.org/xwiki/bin/view/TestReports/ManualTestReportTemplate This report is being copied for each major, minor and intermediary version. Performing a full manual test takes approximately 4-5 days, on one browser. Since we support 4 browsers and 4 databases it would take a very large amount of time to test on everything before every release, since I'm the only one following this manual test report. This is why most of time I perform partial testing on some browsers. Please note that ATM, we stage only one day the artefacts before we release them. Currently our Manual Test Reports is a huge page with a lot of tables. This makes it a little bit hard to test because for each test case, I have to manually add Passed/Failed or No Tested. At the moment there is not enough visibility because we only list the Browsers we support, but we don't give exact information on which testcase has been ran on which browser. I'd like to encourage more users from the community to participate in testing XWiki, and ATM this might be hard because everything is on a single monolithic page. In order to make it as visible as I can, I have the following proposal. The Proposal: Create a dedicated subwiki for QA (maybe qa.xwiki.org), which in future could grow with other QA related things. For the moment, I would like to break the current manual test report in several pages, and several spaces. Each single test would be a wiki page, which will have objects attached to it. The space names will be the features, like WYSIWYG, Administration Application, etc. Also, if this proposal gets voted I'd like in the future to add a new field so we can mark tests that are already automated. This would again, greatly ease my work since I know what is already tested automatically, thus reducing the work needed for manual testing. After brainstorming this idea with Marius and Silvia, we have agreed on the following model: * the MTR will be organized in 2 levels - spaces that will contain feature/testcase suites with pages that hold the individual tests. - pages that will contain individual testcases. Each singular test will also have one or multiple tags attached with the features it belongs to * the following classes to model this: QA.TestClass attached to each page containing a test description (TextArea) - Tells what is the test about e.g: "Tests that images are uploaded fine after a preview." Note that the test name will be the wiki page name and the steps to perform the test will be stored in the document content QA.ExecutionClass (this will contain all information about a test execution) platform (DBList) - linked to QA.PlatformClass browser (DBList) - linked to QA.BrowserClass passed (Boolean) - value to indicate if the test has passed or failed. Absence of an object for a specific platform/browser means not tested jira (String) - link to JIRA issue (if any) tester (User) - user that performed the test date (Date) - date when test was executed QA.PlatformClass (this will contain the version of XE that have been tested/currently being tested) name (String) XE version (String) 4.1.2 QA.BrowserClass (this will contain the browser versions that we support) name (String) Mozilla Firefox version (String) 13 In the same manner we can have this for databases in the future. Final result: - If this is implemented there will be a page with a (custom) lievtable that contains all existing tests. There will be filtering options per feature for example. - There will be also a tag cloud above the livetable (similar with what we have on extensions.xwiki.org) - Marking a test as executed would require 3 clicks maximum (AJAX) right from the LiveTable (Similar to Document/Space Index with Copy/Delete/Rename/Rights), thus improving the speed of manual testing (the "fill in what you tested" part). - Going to the page of an individual test will show all XE versions and browsers on which this has been tested - Easier for everybody to see what has been tested and what not In the future, there will also be a proposal/brainstorming about how to link this with the automated tests. I consider we could be a little bit more visible about this too. Also, there is no visible link between the manual tests and the automated tests. It would be nice if everybody could see for which manual test we have an automated one. I am waiting for your feedback since I am eager to start working on this. Regards, Sorin B. _______________________________________________ devs mailing list [email protected] http://lists.xwiki.org/mailman/listinfo/devs

