This is an automated email from the ASF dual-hosted git repository.
tqchen pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/tvm.git
The following commit(s) were added to refs/heads/main by this push:
new 30b7b1c754 [CI] Upgrade unity image tag to `20240917-153130-9f281758`
(#17410)
30b7b1c754 is described below
commit 30b7b1c7549fbc1277e3a9f5eed73a13f2f0c0ba
Author: Masahiro Hiramori <[email protected]>
AuthorDate: Wed Sep 25 21:52:26 2024 +0900
[CI] Upgrade unity image tag to `20240917-153130-9f281758` (#17410)
* upgrade docker image to `20240917-153130-9f281758`
* fix dynamo test case
* building torch requires c++ 17
* temporary skip jax gpu tests due to XlaRuntimeError
---
ci/jenkins/unity_jenkinsfile.groovy | 8 +++---
src/contrib/msc/plugin/torch_codegen.cc | 2 +-
tests/python/relax/test_frontend_dynamo.py | 2 +-
tests/python/relax/test_frontend_stablehlo.py | 36 ++++++++++++++++++++++++++-
4 files changed, 41 insertions(+), 7 deletions(-)
diff --git a/ci/jenkins/unity_jenkinsfile.groovy
b/ci/jenkins/unity_jenkinsfile.groovy
index 9b4f0009e3..2a7a4fee37 100755
--- a/ci/jenkins/unity_jenkinsfile.groovy
+++ b/ci/jenkins/unity_jenkinsfile.groovy
@@ -30,14 +30,14 @@
import org.jenkinsci.plugins.pipeline.modeldefinition.Utils
// NOTE: these lines are scanned by docker/dev_common.sh. Please update the
regex as needed. -->
-ci_lint = 'tlcpack/ci-lint:20240105-165030-51bdaec6'
-ci_gpu = 'tlcpack/ci-gpu:20240105-165030-51bdaec6'
-ci_cpu = 'tlcpack/ci-cpu:20240105-165030-51bdaec6'
+ci_lint = 'tlcpack/ci_lint:20240917-153130-9f281758'
+ci_gpu = 'tlcpack/ci_gpu:20240917-153130-9f281758'
+ci_cpu = 'tlcpack/ci_cpu:20240917-153130-9f281758'
ci_wasm = 'tlcpack/ci-wasm:v0.72'
ci_i386 = 'tlcpack/ci-i386:v0.75'
ci_qemu = 'tlcpack/ci-qemu:v0.11'
ci_arm = 'tlcpack/ci-arm:v0.08'
-ci_hexagon = 'tlcpack/ci-hexagon:20240105-165030-51bdaec6'
+ci_hexagon = 'tlcpack/ci_hexagon:20240917-153130-9f281758'
// <--- End of regex-scanned config.
// Parameters to allow overriding (in Jenkins UI), the images
diff --git a/src/contrib/msc/plugin/torch_codegen.cc
b/src/contrib/msc/plugin/torch_codegen.cc
index 4b8c24f17b..75471d85db 100644
--- a/src/contrib/msc/plugin/torch_codegen.cc
+++ b/src/contrib/msc/plugin/torch_codegen.cc
@@ -219,7 +219,7 @@ void TorchPluginCodeGen::CodeGenCmake(const
std::set<String>& devices) {
flags.Set("PLUGIN_SUPPORT_TORCH", "");
CodeGenPreCmake(devices, flags);
stack_.line()
- .line("set(CMAKE_CXX_STANDARD 14)")
+ .line("set(CMAKE_CXX_STANDARD 17)")
.line("list(APPEND CMAKE_PREFIX_PATH \"" + config()->torch_prefix +
"\")")
.line("find_package(Torch REQUIRED)");
Array<String> includes, libs;
diff --git a/tests/python/relax/test_frontend_dynamo.py
b/tests/python/relax/test_frontend_dynamo.py
index 21e1d82d28..28215e2e68 100644
--- a/tests/python/relax/test_frontend_dynamo.py
+++ b/tests/python/relax/test_frontend_dynamo.py
@@ -223,7 +223,7 @@ def test_subgraph_capture():
) -> R.Tensor((10,), dtype="float32"):
# block 0
with R.dataflow():
- lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_11,
inp_01)
+ lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_01,
inp_11)
gv1: R.Tensor((10,), dtype="float32") = lv5
R.output(gv1)
return gv1
diff --git a/tests/python/relax/test_frontend_stablehlo.py
b/tests/python/relax/test_frontend_stablehlo.py
index f2d0461dda..667953ab73 100644
--- a/tests/python/relax/test_frontend_stablehlo.py
+++ b/tests/python/relax/test_frontend_stablehlo.py
@@ -196,6 +196,10 @@ def test_add_dynamic():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_unary():
import jax
@@ -229,6 +233,10 @@ def test_unary():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_binary():
import jax
@@ -250,6 +258,10 @@ def test_binary():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_const():
import jax
@@ -260,6 +272,10 @@ def test_const():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_maximum():
import jax
import jax.numpy as jnp
@@ -271,6 +287,10 @@ def test_maximum():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_minimum():
import jax
import jax.numpy as jnp
@@ -282,6 +302,10 @@ def test_minimum():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_reduce():
import jax
import jax.numpy as jnp
@@ -293,6 +317,10 @@ def test_reduce():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_reduce_window():
import jax
from flax import linen as nn
@@ -304,6 +332,10 @@ def test_reduce_window():
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
+# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33
def test_dot_general():
import jax
@@ -314,8 +346,10 @@ def test_dot_general():
check_correctness(jax.jit(fn), input_shapes)
[email protected]()
@tvm.testing.requires_gpu
[email protected](
+ reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN
library initialization failed."
+)
# TODO(yongwww): fix flaky error of "invalid device ordinal"
def test_conv():
import jax