(tvm) branch nightly updated (d1ac1c0202 -> c2d14ae872)

2024-05-13 Thread github-bot
This is an automated email from the ASF dual-hosted git repository. github-bot pushed a change to branch nightly in repository https://gitbox.apache.org/repos/asf/tvm.git from d1ac1c0202 [KVCache] Fix the aux data syncing order of paged KV cache (#16988) add 1d4b9ea5c3 [UnitTest] Use

[PR] [Relay] fixed to make TupleGetItem inherits the previous span [tvm]

2024-05-13 Thread via GitHub
mawnja opened a new pull request, #16996: URL: https://github.com/apache/tvm/pull/16996 [fixed to make TupleGetItem inherits the previous span](https://github.com/apache/tvm/commit/0d14b3cb16c9e0a0087d2c2dc1b82cb6023d93a2) -- This is an automated message from the Apache Git Service. To

(tvm) branch main updated: [Relax][Transform] Handle identical PrimFunc with distinct VDevice (#16959)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git The following commit(s) were added to refs/heads/main by this push: new c2d14ae872 [Relax][Transform] Handle identical

Re: [PR] [Relax][Transform] Handle identical PrimFunc with distinct VDevice [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16959: URL: https://github.com/apache/tvm/pull/16959 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

[PR] [KVCache] Support KVCache decode from forked sequence and pop more tokens [tvm]

2024-05-13 Thread via GitHub
cyx-6 opened a new pull request, #16995: URL: https://github.com/apache/tvm/pull/16995 This PR supports the KVCache to decode from the forked sequence, and pop trailing tokens over multiple blocks. cc: @tqchen @MasterJH5574 -- This is an automated message from the Apache Git

Re: [PR] [Runtime] Allow query of available device memory through DeviceAPI [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on code in PR #16994: URL: https://github.com/apache/tvm/pull/16994#discussion_r1599039598 ## python/tvm/_ffi/runtime_ctypes.py: ## @@ -539,11 +539,25 @@ def total_global_memory(self): Returns --- total_global_memory : int or

Re: [PR] [Cuda] Skip FreeDataSpace when CUDA driver is in inconsistent state [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on code in PR #16980: URL: https://github.com/apache/tvm/pull/16980#discussion_r1599038038 ## src/runtime/cuda/cuda_device_api.cc: ## @@ -142,6 +142,24 @@ class CUDADeviceAPI final : public DeviceAPI { } void FreeDataSpace(Device dev, void* ptr)

(tvm) branch main updated: [QoL][IR] Provide std::hash and std::equal_to for IR Variable types (#16909)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git The following commit(s) were added to refs/heads/main by this push: new 5b5f8d0f77 [QoL][IR] Provide std::hash and

Re: [PR] [QoL][IR] Provide std::hash and std::equal_to for IR Variable types [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16909: URL: https://github.com/apache/tvm/pull/16909 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

Re: [PR] [Disco] Implement `num_workers` property for `disco.Session` [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on PR #16978: URL: https://github.com/apache/tvm/pull/16978#issuecomment-2108697140 Thank you, and conflict resolved! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the

Re: [PR] [Disco] Implement `num_workers` property for `disco.Session` [tvm]

2024-05-13 Thread via GitHub
masahi commented on PR #16978: URL: https://github.com/apache/tvm/pull/16978#issuecomment-2108674980 @Lunderberg Need to resolve the conflict. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to

(tvm) branch main updated: [Unity] Check for transpose and dynamic shape in AdjustMatmulOrder (#16589)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git The following commit(s) were added to refs/heads/main by this push: new 0dfc5f955e [Unity] Check for transpose and dynamic

Re: [PR] [Unity] Check for transpose and dynamic shape in AdjustMatmulOrder [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16589: URL: https://github.com/apache/tvm/pull/16589 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

Re: [PR] [DLight] Check for target in function attributes [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16958: URL: https://github.com/apache/tvm/pull/16958 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

(tvm) branch main updated (29337449db -> eb242ec77b)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a change to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git from 29337449db [Cuda] Skip FreeDataSpace when CUDA driver is in inconsistent state (#16980) add eb242ec77b [DLight]

(tvm) branch main updated (fd820ade5f -> 29337449db)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a change to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git from fd820ade5f [Disco] Expose disco.Session.shutdown through the python API (#16979) add 29337449db [Cuda] Skip

Re: [PR] [Cuda] Skip FreeDataSpace when CUDA driver is in inconsistent state [tvm]

2024-05-13 Thread via GitHub
masahi commented on code in PR #16980: URL: https://github.com/apache/tvm/pull/16980#discussion_r1599001463 ## src/runtime/cuda/cuda_device_api.cc: ## @@ -142,6 +142,24 @@ class CUDADeviceAPI final : public DeviceAPI { } void FreeDataSpace(Device dev, void* ptr) final {

Re: [PR] [Cuda] Skip FreeDataSpace when CUDA driver is in inconsistent state [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16980: URL: https://github.com/apache/tvm/pull/16980 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

(tvm) branch main updated: [Disco] Expose disco.Session.shutdown through the python API (#16979)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git The following commit(s) were added to refs/heads/main by this push: new fd820ade5f [Disco] Expose disco.Session.shutdown

Re: [PR] [Disco] Expose disco.Session.shutdown through the python API [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16979: URL: https://github.com/apache/tvm/pull/16979 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

(tvm) branch main updated (d1ac1c0202 -> 1d4b9ea5c3)

2024-05-13 Thread masahi
This is an automated email from the ASF dual-hosted git repository. masahi pushed a change to branch main in repository https://gitbox.apache.org/repos/asf/tvm.git from d1ac1c0202 [KVCache] Fix the aux data syncing order of paged KV cache (#16988) add 1d4b9ea5c3 [UnitTest] Use

Re: [PR] [UnitTest] Use pytest's scope='session' for tvm.testing.parameter [tvm]

2024-05-13 Thread via GitHub
masahi merged PR #16930: URL: https://github.com/apache/tvm/pull/16930 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail:

Re: [PR] [Runtime] Allow query of available device memory through DeviceAPI [tvm]

2024-05-13 Thread via GitHub
masahi commented on code in PR #16994: URL: https://github.com/apache/tvm/pull/16994#discussion_r1598985475 ## python/tvm/_ffi/runtime_ctypes.py: ## @@ -539,11 +539,25 @@ def total_global_memory(self): Returns --- total_global_memory : int or None

Re: [PR] [Disco] Implement `num_workers` property for `disco.Session` [tvm]

2024-05-13 Thread via GitHub
Lunderberg closed pull request #16990: [Disco] Implement `num_workers` property for `disco.Session` URL: https://github.com/apache/tvm/pull/16990 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the

Re: [PR] [Disco] Implement `num_workers` property for `disco.Session` [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on PR #16990: URL: https://github.com/apache/tvm/pull/16990#issuecomment-2108326695 Closing this PR, as it is identical to https://github.com/apache/tvm/pull/16978. I think that means I have too many open PRs, and any help in reviewing them would be appreciated. --

[PR] [Runtime] Allow query of available device memory through DeviceAPI [tvm]

2024-05-13 Thread via GitHub
Lunderberg opened a new pull request, #16994: URL: https://github.com/apache/tvm/pull/16994 Prior to this commit, the total device memory could be queried through the `DeviceAPI` interface, but the currently available device memory could not. This functionality may be useful for

[PR] [Disco] Allow allocation that only exists on worker0 [tvm]

2024-05-13 Thread via GitHub
Lunderberg opened a new pull request, #16993: URL: https://github.com/apache/tvm/pull/16993 The `disco.Session.scatter_from_worker0` function expects a `DRef` which an `NDArray` on worker 0, and `NullOpt` on all other workers. Prior to this commit, there was no method in the

Re: [PR] [SVE] Add support for representing and creating buffer-level predicates [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on PR #16966: URL: https://github.com/apache/tvm/pull/16966#issuecomment-2108122470 No problem, and thank you on the revisions! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go

Re: [PR] [SVE] Add support for representing and creating buffer-level predicates [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on code in PR #16966: URL: https://github.com/apache/tvm/pull/16966#discussion_r1598726836 ## src/target/llvm/codegen_llvm.cc: ## @@ -1768,11 +1774,17 @@ llvm::Value* CodeGenLLVM::VisitExpr_(const BufferLoadNode* op) { std::vector loads; - auto

Re: [PR] [SVE] Add support for representing and creating buffer-level predicates [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on code in PR #16966: URL: https://github.com/apache/tvm/pull/16966#discussion_r1598722953 ## python/tvm/tir/buffer.py: ## @@ -141,6 +141,57 @@ def vstore(self, begin, value): begin = (begin,) if isinstance(begin, (int, PrimExpr)) else begin

Re: [PR] [SVE] Add support for representing and creating buffer-level predicates [tvm]

2024-05-13 Thread via GitHub
Lunderberg commented on code in PR #16966: URL: https://github.com/apache/tvm/pull/16966#discussion_r1598715370 ## src/tir/transforms/vectorize_loop.cc: ## @@ -72,6 +72,126 @@ inline PrimExpr BroadcastTo(PrimExpr e, int lanes, bool is_scalable) { return Broadcast(e,

[PR] [Bugfix][Disco] Handle NDArray larger than OS buffer for pipe [tvm]

2024-05-13 Thread via GitHub
Lunderberg opened a new pull request, #16992: URL: https://github.com/apache/tvm/pull/16992 Prior to this commit, using `disco.Session` methods to transfer `NDArray` instances to workers could raise an exception if the `NDArray` is larger than the buffer allocated by the OS for the

Re: [PR] [SVE] Add support for representing and creating buffer-level predicates [tvm]

2024-05-13 Thread via GitHub
lhutton1 commented on code in PR #16966: URL: https://github.com/apache/tvm/pull/16966#discussion_r1598578740 ## src/tir/transforms/vectorize_loop.cc: ## @@ -72,6 +72,126 @@ inline PrimExpr BroadcastTo(PrimExpr e, int lanes, bool is_scalable) { return Broadcast(e,

[I] [Bug] Segmentation Fault when Running Relay Transformations Iteratively [tvm]

2024-05-13 Thread via GitHub
Jupiterghy opened a new issue, #16991: URL: https://github.com/apache/tvm/issues/16991 Executing a sequence of Relay transformations iteratively using a custom script results in a segmentation fault (Segmentation fault (core dumped)). This issue seems to be related to the quantity of

[PR] [Disco] Implement `num_workers` property for `disco.Session` [tvm]

2024-05-13 Thread via GitHub
Lunderberg opened a new pull request, #16990: URL: https://github.com/apache/tvm/pull/16990 Prior to this commit, while the `num_workers` argument was provided to the `disco.Session` object, it could not be determined from an existing `disco.Session` object. As a result, functions that