This is an automated email from the ASF dual-hosted git repository.

mehrdadh pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/tvm.git


The following commit(s) were added to refs/heads/main by this push:
     new a9ece3d48b [hexagon][testing] refactor benchmark-table code (#11400)
a9ece3d48b is described below

commit a9ece3d48b40b4ecc496eaf5b2f831a0e2817620
Author: Christian Convey <[email protected]>
AuthorDate: Thu May 26 12:02:31 2022 -0400

    [hexagon][testing] refactor benchmark-table code (#11400)
    
    Generalize the benchmark-table code to support arbitrary
    independent values. This supports future changes to the benchmark
    code.
---
 .../contrib/test_hexagon/benchmark_hexagon.py      | 149 ++++++---------------
 .../python/contrib/test_hexagon/benchmark_util.py  | 141 +++++++++++++++++++
 2 files changed, 180 insertions(+), 110 deletions(-)

diff --git a/tests/python/contrib/test_hexagon/benchmark_hexagon.py 
b/tests/python/contrib/test_hexagon/benchmark_hexagon.py
index 979bd11170..2a1d6796e7 100644
--- a/tests/python/contrib/test_hexagon/benchmark_hexagon.py
+++ b/tests/python/contrib/test_hexagon/benchmark_hexagon.py
@@ -17,17 +17,16 @@
 
 import os
 import os.path
-import pathlib
 import sys
 import pytest
 import numpy as np
 import logging
 import tempfile
-import csv
 
 import tvm.testing
 from tvm import te
 from tvm.contrib.hexagon.build import HexagonLauncherRPC
+from .benchmark_util import BenchmarksTable
 
 RPC_SERVER_PORT = 7070
 
@@ -58,112 +57,22 @@ def test_elemwise_add(hexagon_launcher: 
HexagonLauncherRPC):
     print("-" * 80)
     print()
 
-    # TODO: We should move this into a separate test fixture, to make it 
easier to write
-    # additional benchmarking functions.  We'd just need to generalize the 
assumptions regarding
-    # the particular fields being tracked as independent variables.
-    class benchmark_results_collection:
-        def __init__(self):
-            self.row_dicts_ = []
-
-        def num_failures(self):
-            num = 0
-            for d in self.row_dicts_:
-                if d["status"] == "FAIL":
-                    num += 1
-            return num
-
-        def num_skips(self):
-            num = 0
-            for d in self.row_dicts_:
-                if d["status"] == "SKIP":
-                    num += 1
-            return num
-
-        def record_success(
-            self, dtype, sched_type, mem_scope, num_vecs_per_tensor, 
benchmark_result
-        ):
-            median_usec = benchmark_result.median * 1000000
-            min_usec = benchmark_result.min * 1000000
-            max_usec = benchmark_result.max * 1000000
-
-            self.row_dicts_.append(
-                {
-                    "dtype": dtype,
-                    "sched_type": sched_type,
-                    "mem_scope": mem_scope,
-                    "num_vecs_per_tensor": num_vecs_per_tensor,
-                    "status": "OK",
-                    "median(µsec)": f"{median_usec:.3}",
-                    "min(µsec)": f"{min_usec:.3}",
-                    "max(µsec)": f"{max_usec:.3}",
-                }
-            )
-
-        def record_failure(self, dtype, sched_type, mem_scope, 
num_vecs_per_tensor, error_text):
-            self.row_dicts_.append(
-                {
-                    "dtype": dtype,
-                    "sched_type": sched_type,
-                    "mem_scope": mem_scope,
-                    "num_vecs_per_tensor": num_vecs_per_tensor,
-                    "status": "FAIL",
-                    "comment": error_text,
-                }
-            )
-
-        def record_skip(self, dtype, sched_type, mem_scope, 
num_vecs_per_tensor, comment_text):
-            self.row_dicts_.append(
-                {
-                    "dtype": dtype,
-                    "sched_type": sched_type,
-                    "mem_scope": mem_scope,
-                    "num_vecs_per_tensor": num_vecs_per_tensor,
-                    "status": "SKIP",
-                    "comment": comment_text,
-                }
-            )
-
-        def dump(self, f):
-            csv.register_dialect(
-                "benchmarks",
-                delimiter="\t",
-                quotechar='"',
-                quoting=csv.QUOTE_MINIMAL,
-            )
-
-            fieldnames = [
-                "dtype",
-                "sched_type",
-                "mem_scope",
-                "num_vecs_per_tensor",
-                "status",
-                "median(µsec)",
-                "min(µsec)",
-                "max(µsec)",
-                "comment",
-            ]
-
-            writer = csv.DictWriter(f, fieldnames, dialect="benchmarks", 
restval="")
-
-            writer.writeheader()
-            for d in self.row_dicts_:
-                writer.writerow(d)
-
-    br = benchmark_results_collection()
+    bt = BenchmarksTable()
 
     # Create and benchmark a single primfunc.
-    # If an unexpected problem occurs, raise an exception.  Otherwise add a 
row of output to 'br'.
+    # If an unexpected problem occurs, raise an exception.  Otherwise add a 
row of output to 'bt'.
     def test_one_config(dtype, sched_type, mem_scope, num_vectors_per_tensor):
         version_name = 
f"dtype:{dtype}-schedtype:{sched_type}-memscope:{mem_scope}-numvecs:{num_vectors_per_tensor}"
+        print()
         print(f"CONFIGURATION: {version_name}")
 
         if num_vectors_per_tensor == 2048 and mem_scope == "global.vtcm":
-            br.record_skip(
-                dtype,
-                sched_type,
-                mem_scope,
-                num_vectors_per_tensor,
-                f"Expect to exceed VTCM budget.",
+            bt.record_skip(
+                dtype=dtype,
+                sched_type=sched_type,
+                mem_scope=mem_scope,
+                num_vectors_per_tensor=num_vectors_per_tensor,
+                comments="Expect to exceed VTCM budget.",
             )
             return
 
@@ -255,25 +164,45 @@ def test_elemwise_add(hexagon_launcher: 
HexagonLauncherRPC):
                     timer = mod.time_evaluator("elemwise_add", sess.device, 
number=10, repeat=1)
                     timing_result = timer(A_data, B_data, C_data)
 
-                    print("TIMING RESULT: {}".format(timing_result))
-
                     # Verify that the computation actually happened, and 
produced the correct result.
                     result = C_data.numpy()
                     tvm.testing.assert_allclose(host_numpy_C_data_expected, 
result)
 
-                    br.record_success(
-                        dtype, sched_type, mem_scope, num_vectors_per_tensor, 
timing_result
+                    bt.record_success(
+                        timing_result,
+                        dtype=dtype,
+                        sched_type=sched_type,
+                        mem_scope=mem_scope,
+                        num_vectors_per_tensor=num_vectors_per_tensor,
                     )
 
             except Exception as err:
                 f.write("ERROR:\n")
                 f.write("{}\n".format(err))
-                br.record_failure(
-                    dtype, sched_type, mem_scope, num_vectors_per_tensor, 
f"See {report_path}"
+                bt.record_fail(
+                    dtype=dtype,
+                    sched_type=sched_type,
+                    mem_scope=mem_scope,
+                    num_vectors_per_tensor=num_vectors_per_tensor,
+                    comments=f"See {report_path}",
                 )
 
     # 
-----------------------------------------------------------------------------------------------
 
+    csv_column_order = [
+        "dtype",
+        "sched_type",
+        "mem_scope",
+        "num_vectors_per_tensor",
+        "row_status",
+        "timings_min_usecs",
+        "timings_max_usecs",
+        "timings_median_usecs",
+        "timings_mean_usecs",
+        "timings_stddev_usecs",
+        "comments",
+    ]
+
     # Hexagon v69 allows more dtypes, but we're sticking with v68 for now.
     for dtype in [
         "int8",
@@ -300,7 +229,7 @@ def test_elemwise_add(hexagon_launcher: HexagonLauncherRPC):
                     test_one_config(dtype, sched_type, mem_scope, 
num_vectors_per_tensor)
 
                     # Report our progress.
-                    br.dump(sys.stdout)
+                    bt.print_csv(sys.stdout, csv_column_order)
 
     print("-" * 80)
     print(f"OUTPUT DIRECTORY: {host_output_dir}")
@@ -309,8 +238,8 @@ def test_elemwise_add(hexagon_launcher: HexagonLauncherRPC):
 
     tabular_output_filename = os.path.join(host_output_dir, 
"benchmark-results.csv")
     with open(tabular_output_filename, "w") as csv_file:
-        br.dump(csv_file)
+        bt.print_csv(csv_file, csv_column_order)
     print(f"BENCHMARK RESULTS FILE: {tabular_output_filename}")
 
-    if br.num_failures() > 0:
+    if bt.has_fail() > 0:
         pytest.fail("At least one benchmark configuration failed", 
pytrace=False)
diff --git a/tests/python/contrib/test_hexagon/benchmark_util.py 
b/tests/python/contrib/test_hexagon/benchmark_util.py
new file mode 100644
index 0000000000..5a75e9a6e8
--- /dev/null
+++ b/tests/python/contrib/test_hexagon/benchmark_util.py
@@ -0,0 +1,141 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+import csv
+
+
+class BenchmarksTable:
+    """
+    Stores/reports the result of benchmark runs.
+
+    Each line item has a status: success, fail, or skip.
+
+    Each 'success' line item must include benchmark data,
+    in the form provided by TVM's `time_evaluator` mechanism.
+
+    Each line item may also specify values for any subset of
+    the columns provided to the table's construstor.
+    """
+
+    BUILTIN_COLUMN_NAMES = set(
+        [
+            "row_status",
+            "timings_min_usecs",
+            "timings_max_usecs",
+            "timings_median_usecs",
+            "timings_mean_usecs",
+            "timings_stddev_usecs",
+        ]
+    )
+
+    def __init__(self):
+        self._line_items = []
+
+    def validate_user_supplied_kwargs(self, kwarg_dict):
+        name_conflicts = 
set(kwarg_dict).intersection(self.BUILTIN_COLUMN_NAMES)
+
+        if name_conflicts:
+            name_list = ", ".join(name_conflicts)
+            raise Exception(f"Attempting to supply values for built-in column 
names: {name_list}")
+
+    def record_success(self, timings, **kwargs):
+        """
+        `timings` : Assumed to have the structure and meaning of
+          the timing results provided by TVM's `time_evaluator`
+          mechanism.
+
+        `kwargs` : Optional values for any of the other columns
+          defined for this benchmark table.
+        """
+        self.validate_user_supplied_kwargs(kwargs)
+        line_item = kwargs
+
+        line_item["row_status"] = "SUCCESS"
+
+        line_item["timings_min_usecs"] = timings.min * 1000000
+        line_item["timings_max_usecs"] = timings.max * 1000000
+        line_item["timings_median_usecs"] = timings.median * 1000000
+        line_item["timings_stddev_usecs"] = timings.std * 1000000
+        line_item["timings_mean_usecs"] = timings.mean * 1000000
+
+        self._line_items.append(line_item)
+
+    def record_skip(self, **kwargs):
+        self.validate_user_supplied_kwargs(kwargs)
+
+        line_item = dict(kwargs)
+        line_item["row_status"] = "SKIP"
+        self._line_items.append(line_item)
+
+    def record_fail(self, **kwargs):
+        self.validate_user_supplied_kwargs(kwargs)
+
+        line_item = dict(kwargs)
+        line_item["row_status"] = "FAIL"
+        self._line_items.append(line_item)
+
+    def has_fail(self):
+        """
+        Returns True if the table contains at least one 'fail' line item,
+        otherwise returns False.
+        """
+        return any(item["row_status"] == "FAIL" for item in self._line_items)
+
+    def print_csv(self, f, column_name_order, timing_decimal_places=3):
+        """
+        Print the benchmark results as a csv.
+
+        `f` : The output stream.
+
+        `column_name_order`: an iterable sequence of column names, indicating 
the
+           left-to-right ordering of columns in the CSV output.
+
+           The CSV output will contain only those columns that are mentioned in
+           this list.
+
+        `timing_decimal_places`: for the numeric timing values, this is the
+           number of decimal places to provide in the printed output.
+           For example, a value of 3 is equivalent to the Python formatting 
string
+           `'{:.3f}'`
+        """
+        writer = csv.DictWriter(
+            f, column_name_order, dialect="excel-tab", restval="", 
extrasaction="ignore"
+        )
+
+        writer.writeheader()
+
+        for line_item_dict in self._line_items:
+            # Use a copy of the line-item dictionary, because we might do some 
modifications
+            # for the sake of rendering...
+            csv_line_dict = dict(line_item_dict)
+
+            for col_name in [
+                "timings_min_usecs",
+                "timings_max_usecs",
+                "timings_median_usecs",
+                "timings_stddev_usecs",
+                "timings_mean_usecs",
+            ]:
+                if col_name in csv_line_dict:
+                    old_value = csv_line_dict[col_name]
+                    assert isinstance(
+                        old_value, float
+                    ), f"Formatting code assumes that column {col_name} is 
some col_nameind of float, but its actual type is {type(old_value)}"
+                    str_value = f"{old_value:>0.{timing_decimal_places}f}"
+                    csv_line_dict[col_name] = str_value
+
+            writer.writerow(csv_line_dict)

Reply via email to