Paul Rogers created IMPALA-8055: ----------------------------------- Summary: run-tests.py reports tests as passed even if the did not Key: IMPALA-8055 URL: https://issues.apache.org/jira/browse/IMPALA-8055 Project: IMPALA Issue Type: Bug Components: Infrastructure Affects Versions: Impala 3.1.0 Reporter: Paul Rogers
Been mucking about with the EXPLAIN output format which required rebasing a bunch of tests on the new format. PlannerTest is fine: it clearly fails when the expected ".test" files don't match the new "actual" files. When run on Jenkins in "pre-review" mode, the build does fail if a Python end-to-end test fails. But, the job seems to give up at that point, not running other tests and finding more problems. (There were three separate test cases that needed fixing; took multiple runs to find them.) When run on my dev box, I get the following (highly abbreviated) output: {noformat} '| in pipelines: 00(GETNEXT)' != '| row-size=402B cardinality=5.76M' ... [gw3] PASSED metadata/test_explain.py::TestExplain::test_explain_level0[protocol: beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'debug_action': None, 'exec_single_node_rows_threshold': 0} | table_format: text/none] ... ==== 6 passed in 68.63 seconds ===== {noformat} I've learned that "passed" means "maybe failed" and to go back and inspect the actual output to figure out if the test did, indeed, fail. I suspect "passed" means "didn't crash" rather than "tests worked." Would be very helpful to plumb the failure through to the summary line so it said "3 passed, 3 failed" or whatever. Would be a huge time-saver. -- This message was sent by Atlassian JIRA (v7.6.3#76005)