CLIMATE-581 - Fix minor trailing/leading whitespace issues
Project: http://git-wip-us.apache.org/repos/asf/climate/repo Commit: http://git-wip-us.apache.org/repos/asf/climate/commit/9167f1a5 Tree: http://git-wip-us.apache.org/repos/asf/climate/tree/9167f1a5 Diff: http://git-wip-us.apache.org/repos/asf/climate/diff/9167f1a5 Branch: refs/heads/master Commit: 9167f1a5c8366ecda77693a61c49f345ce2c41d8 Parents: f83e4be Author: Michael Joyce <[email protected]> Authored: Wed Mar 25 11:33:22 2015 -0700 Committer: Michael Joyce <[email protected]> Committed: Wed Mar 25 11:33:22 2015 -0700 ---------------------------------------------------------------------- ocw-config-runner/configuration_writer.py | 6 +++--- ocw-config-runner/tests/test_config_writer.py | 12 ++++++------ 2 files changed, 9 insertions(+), 9 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/climate/blob/9167f1a5/ocw-config-runner/configuration_writer.py ---------------------------------------------------------------------- diff --git a/ocw-config-runner/configuration_writer.py b/ocw-config-runner/configuration_writer.py index c78f46f..f7fb1c9 100644 --- a/ocw-config-runner/configuration_writer.py +++ b/ocw-config-runner/configuration_writer.py @@ -25,7 +25,7 @@ logger = logging.getLogger(__name__) def export_evaluation_to_config(evaluation, file_path='./exported_eval.yaml'): ''' Export an evaluation to a config file - + :param evaluation: The evaluation object to export. :type evaluation: :class:`evaluation.Evaluation` @@ -42,7 +42,7 @@ def export_evaluation_to_config(evaluation, file_path='./exported_eval.yaml'): def generate_dataset_information(evaluation): ''' Generate dataset config file output for a given Evaluation object. - + :param evaluation: The evaluation object from which to extract metrics. :type evaluation: :class:`evaluation.Evaluation` @@ -124,7 +124,7 @@ def generate_evaluation_information(evaluation): configuration information. It's possible that you will encounter a scenario where the guessed values are not what you want/expect. Please double check the output before blinding trusting what this generates. - + :param evaluation: The evaluation object from which to extract metrics. :type evaluation: :class:`evaluation.Evaluation` http://git-wip-us.apache.org/repos/asf/climate/blob/9167f1a5/ocw-config-runner/tests/test_config_writer.py ---------------------------------------------------------------------- diff --git a/ocw-config-runner/tests/test_config_writer.py b/ocw-config-runner/tests/test_config_writer.py index 2d6a9f6..97b7e01 100644 --- a/ocw-config-runner/tests/test_config_writer.py +++ b/ocw-config-runner/tests/test_config_writer.py @@ -293,7 +293,7 @@ class TestDatasetExportFromEvaluation(unittest.TestCase): 'time_name': 'a time name', 'elevation_index': 2 } - + self.rcmed_origin = { 'source': 'rcmed', 'dataset_id': 4, @@ -332,7 +332,7 @@ class TestDatasetExportFromEvaluation(unittest.TestCase): name=self.name, origin=self.rcmed_origin ) - + self.esgf_ds = Dataset( self.lats, self.lons, @@ -589,7 +589,7 @@ class FullExportTest(unittest.TestCase): 'time_name': 'a time name', 'elevation_index': 2 } - + self.rcmed_origin = { 'source': 'rcmed', 'dataset_id': 4, @@ -628,7 +628,7 @@ class FullExportTest(unittest.TestCase): name=self.name, origin=self.rcmed_origin ) - + self.esgf_ds = Dataset( self.lats, self.lons, @@ -690,7 +690,7 @@ class FullExportTest(unittest.TestCase): self.assertTrue(metric.__class__.__name__ in data['metrics']) total_eval_metrics = ( - len(self.evaluation.metrics) + + len(self.evaluation.metrics) + len(self.evaluation.unary_metrics) ) @@ -710,7 +710,7 @@ class FullExportTest(unittest.TestCase): self.assertTrue('targets' in data['datasets']) self.assertAlmostEqual( - writer.generate_dataset_information(self.evaluation), + writer.generate_dataset_information(self.evaluation), data['datasets'] )
