This is an automated email from the ASF dual-hosted git repository.

tqchen pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/tvm.git


The following commit(s) were added to refs/heads/main by this push:
     new 30b7b1c754 [CI] Upgrade unity image tag to `20240917-153130-9f281758` 
(#17410)
30b7b1c754 is described below

commit 30b7b1c7549fbc1277e3a9f5eed73a13f2f0c0ba
Author: Masahiro Hiramori <[email protected]>
AuthorDate: Wed Sep 25 21:52:26 2024 +0900

    [CI] Upgrade unity image tag to `20240917-153130-9f281758` (#17410)
    
    * upgrade docker image to `20240917-153130-9f281758`
    
    * fix dynamo test case
    
    * building torch requires c++ 17
    
    * temporary skip jax gpu tests due to XlaRuntimeError
---
 ci/jenkins/unity_jenkinsfile.groovy           |  8 +++---
 src/contrib/msc/plugin/torch_codegen.cc       |  2 +-
 tests/python/relax/test_frontend_dynamo.py    |  2 +-
 tests/python/relax/test_frontend_stablehlo.py | 36 ++++++++++++++++++++++++++-
 4 files changed, 41 insertions(+), 7 deletions(-)

diff --git a/ci/jenkins/unity_jenkinsfile.groovy 
b/ci/jenkins/unity_jenkinsfile.groovy
index 9b4f0009e3..2a7a4fee37 100755
--- a/ci/jenkins/unity_jenkinsfile.groovy
+++ b/ci/jenkins/unity_jenkinsfile.groovy
@@ -30,14 +30,14 @@
 import org.jenkinsci.plugins.pipeline.modeldefinition.Utils
 
 // NOTE: these lines are scanned by docker/dev_common.sh. Please update the 
regex as needed. -->
-ci_lint = 'tlcpack/ci-lint:20240105-165030-51bdaec6'
-ci_gpu = 'tlcpack/ci-gpu:20240105-165030-51bdaec6'
-ci_cpu = 'tlcpack/ci-cpu:20240105-165030-51bdaec6'
+ci_lint = 'tlcpack/ci_lint:20240917-153130-9f281758'
+ci_gpu = 'tlcpack/ci_gpu:20240917-153130-9f281758'
+ci_cpu = 'tlcpack/ci_cpu:20240917-153130-9f281758'
 ci_wasm = 'tlcpack/ci-wasm:v0.72'
 ci_i386 = 'tlcpack/ci-i386:v0.75'
 ci_qemu = 'tlcpack/ci-qemu:v0.11'
 ci_arm = 'tlcpack/ci-arm:v0.08'
-ci_hexagon = 'tlcpack/ci-hexagon:20240105-165030-51bdaec6'
+ci_hexagon = 'tlcpack/ci_hexagon:20240917-153130-9f281758'
 // <--- End of regex-scanned config.
 
 // Parameters to allow overriding (in Jenkins UI), the images
diff --git a/src/contrib/msc/plugin/torch_codegen.cc 
b/src/contrib/msc/plugin/torch_codegen.cc
index 4b8c24f17b..75471d85db 100644
--- a/src/contrib/msc/plugin/torch_codegen.cc
+++ b/src/contrib/msc/plugin/torch_codegen.cc
@@ -219,7 +219,7 @@ void TorchPluginCodeGen::CodeGenCmake(const 
std::set<String>& devices) {
   flags.Set("PLUGIN_SUPPORT_TORCH", "");
   CodeGenPreCmake(devices, flags);
   stack_.line()
-      .line("set(CMAKE_CXX_STANDARD 14)")
+      .line("set(CMAKE_CXX_STANDARD 17)")
       .line("list(APPEND CMAKE_PREFIX_PATH \"" + config()->torch_prefix + 
"\")")
       .line("find_package(Torch REQUIRED)");
   Array<String> includes, libs;
diff --git a/tests/python/relax/test_frontend_dynamo.py 
b/tests/python/relax/test_frontend_dynamo.py
index 21e1d82d28..28215e2e68 100644
--- a/tests/python/relax/test_frontend_dynamo.py
+++ b/tests/python/relax/test_frontend_dynamo.py
@@ -223,7 +223,7 @@ def test_subgraph_capture():
         ) -> R.Tensor((10,), dtype="float32"):
             # block 0
             with R.dataflow():
-                lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_11, 
inp_01)
+                lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_01, 
inp_11)
                 gv1: R.Tensor((10,), dtype="float32") = lv5
                 R.output(gv1)
             return gv1
diff --git a/tests/python/relax/test_frontend_stablehlo.py 
b/tests/python/relax/test_frontend_stablehlo.py
index f2d0461dda..667953ab73 100644
--- a/tests/python/relax/test_frontend_stablehlo.py
+++ b/tests/python/relax/test_frontend_stablehlo.py
@@ -196,6 +196,10 @@ def test_add_dynamic():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_unary():
     import jax
 
@@ -229,6 +233,10 @@ def test_unary():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_binary():
     import jax
 
@@ -250,6 +258,10 @@ def test_binary():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_const():
     import jax
 
@@ -260,6 +272,10 @@ def test_const():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_maximum():
     import jax
     import jax.numpy as jnp
@@ -271,6 +287,10 @@ def test_maximum():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_minimum():
     import jax
     import jax.numpy as jnp
@@ -282,6 +302,10 @@ def test_minimum():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_reduce():
     import jax
     import jax.numpy as jnp
@@ -293,6 +317,10 @@ def test_reduce():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_reduce_window():
     import jax
     from flax import linen as nn
@@ -304,6 +332,10 @@ def test_reduce_window():
 
 
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
 def test_dot_general():
     import jax
 
@@ -314,8 +346,10 @@ def test_dot_general():
     check_correctness(jax.jit(fn), input_shapes)
 
 
[email protected]()
 @tvm.testing.requires_gpu
[email protected](
+    reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN 
library initialization failed."
+)
 # TODO(yongwww): fix flaky error of "invalid device ordinal"
 def test_conv():
     import jax

Reply via email to