This is an automated email from the ASF dual-hosted git repository.

tqchen pushed a commit to branch refactor-s1
in repository https://gitbox.apache.org/repos/asf/tvm.git


The following commit(s) were added to refs/heads/refactor-s1 by this push:
     new 6fdba43942 fix lin and MS
6fdba43942 is described below

commit 6fdba439427cf0aea9cb6660d7be90d4f9cd8c58
Author: tqchen <[email protected]>
AuthorDate: Thu Apr 17 20:13:36 2025 -0400

    fix lin and MS
---
 .../python/contrib/test_hexagon/test_dma_builtin.py | 21 +++------------------
 tests/python/contrib/test_msc/test_plugin.py        |  4 +++-
 .../meta_schedule/test_meta_schedule_trace_apply.py |  4 ++--
 3 files changed, 8 insertions(+), 21 deletions(-)

diff --git a/tests/python/contrib/test_hexagon/test_dma_builtin.py 
b/tests/python/contrib/test_hexagon/test_dma_builtin.py
index 1e818dd00a..479b680065 100644
--- a/tests/python/contrib/test_hexagon/test_dma_builtin.py
+++ b/tests/python/contrib/test_hexagon/test_dma_builtin.py
@@ -65,12 +65,7 @@ class Module_1D:
             dtype=data_type,
             storage_scope="global.vtcm",
         )
-        a: R.Tensor(
-            [
-                12800,
-            ],
-            dtype=data_type,
-        ) = R.vm.alloc_tensor(
+        a: R.Tensor([12800,], dtype=data_type,) = R.vm.alloc_tensor(
             vtcm_obj,
             offset=0,
             shape=R.shape(
@@ -85,12 +80,7 @@ class Module_1D:
             [x, a, 0, True],
             sinfo_args=[],
         )
-        b: R.Tensor(
-            [
-                12800,
-            ],
-            dtype=data_type,
-        ) = R.vm.alloc_tensor(
+        b: R.Tensor([12800,], dtype=data_type,) = R.vm.alloc_tensor(
             vtcm_obj,
             offset=12800 * 4,
             shape=R.shape(
@@ -105,12 +95,7 @@ class Module_1D:
             [y, b, 1, True],
             sinfo_args=[],
         )
-        c: R.Tensor(
-            [
-                12800,
-            ],
-            dtype=data_type,
-        ) = R.vm.alloc_tensor(
+        c: R.Tensor([12800,], dtype=data_type,) = R.vm.alloc_tensor(
             vtcm_obj,
             offset=2 * 12800 * 4,
             shape=R.shape(
diff --git a/tests/python/contrib/test_msc/test_plugin.py 
b/tests/python/contrib/test_msc/test_plugin.py
index 1c78dbfdd1..30eb6dd067 100644
--- a/tests/python/contrib/test_msc/test_plugin.py
+++ b/tests/python/contrib/test_msc/test_plugin.py
@@ -322,7 +322,9 @@ def _test_with_manager(plugins, compile_type, 
expected_info):
     ), "Model info {} mismatch with expected {}".format(model_info, 
expected_info)
 
 
[email protected](reason="skip the test because plugin needs to include ffi 
folder, can be re-enabled")
[email protected](
+    reason="skip the test because plugin needs to include ffi folder, can be 
re-enabled"
+)
 def test_plugin():
     """Test the plugins"""
 
diff --git a/tests/python/meta_schedule/test_meta_schedule_trace_apply.py 
b/tests/python/meta_schedule/test_meta_schedule_trace_apply.py
index 3f6f2e1a65..c3a76e101f 100644
--- a/tests/python/meta_schedule/test_meta_schedule_trace_apply.py
+++ b/tests/python/meta_schedule/test_meta_schedule_trace_apply.py
@@ -105,7 +105,7 @@ class DenseAdd_scheduled_cpu:
                 v0, v1 = T.axis.remap("SS", [ax0, ax1])
                 T.reads(p1[v0, v1])
                 T.writes(p1_global[v0 // 64, v1, v0 % 64])
-                T.block_attr({"meta_schedule.layout_rewrite_preproc": 1})
+                T.block_attr({"meta_schedule.layout_rewrite_preproc": True})
                 p1_global[v0 // 64, v1, v0 % 64] = p1[v0, v1]
         for i0_0_i1_0_fused_fused in T.parallel(4):
             for i0_1, i1_1 in T.grid(8, 1):
@@ -184,7 +184,7 @@ class DenseAdd_cpu_no_write_cache:
                 v0, v1 = T.axis.remap("SS", [ax0, ax1])
                 T.reads(p1[v0, v1])
                 T.writes(p1_global[v1 // 16, v0 // 32, v1 % 16, v0 % 32])
-                T.block_attr({"meta_schedule.layout_rewrite_preproc":1})
+                T.block_attr({"meta_schedule.layout_rewrite_preproc": True})
                 p1_global[v1 // 16, v0 // 32, v1 % 16, v0 % 32] = p1[v0, v1]
         for i0_0_i1_0_i0_1_i1_1_fused in T.parallel(16, 
annotations={"pragma_auto_unroll_max_step":16, "pragma_unroll_explicit":1}):
             for i0_2_init, i1_2_init, i0_3_init in T.grid(4, 4, 2):

Reply via email to