driazati opened a new issue, #13284:
URL: https://github.com/apache/tvm/issues/13284

   Seen here and elsewhere: 
https://ci.tlcpack.ai/blue/organizations/jenkins/tvm/detail/main/4628/pipeline
   
   ```
   [2022-11-02T15:27:34.310Z]   File "tvm/_ffi/_cython/./packed_func.pxi", line 
56, in tvm._ffi._cy3.core.tvm_callback
   [2022-11-02T15:27:34.310Z]   File 
"/workspace/python/tvm/autotvm/measure/measure_methods.py", line 871, in 
verify_pass
   [2022-11-02T15:27:34.310Z]     raise InstantiationError("Skipped because of 
invalid gpu kernel")
   [2022-11-02T15:27:34.310Z] tvm.autotvm.task.space.InstantiationError: 
Skipped because of invalid gpu kernel
   [2022-11-02T15:27:34.310Z] 
   [2022-11-02T15:27:34.310Z] Traceback (most recent call last):
   [2022-11-02T15:27:34.310Z]   24: TVMFuncCall
   [2022-11-02T15:27:34.310Z]         at ../src/runtime/c_runtime_api.cc:477
   [2022-11-02T15:27:34.310Z]   23: 
tvm::runtime::PackedFuncObj::CallPacked(tvm::runtime::TVMArgs, 
tvm::runtime::TVMRetValue*) const
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1217
   [2022-11-02T15:27:34.310Z]   22: Call
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1213
   [2022-11-02T15:27:34.310Z]   21: operator()
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1731
   [2022-11-02T15:27:34.310Z]   20: unpack_call<tvm::IRModule, 5, 
tvm::<lambda(tvm::te::Schedule, const 
tvm::runtime::Array<tvm::runtime::ObjectRef>&, const tvm::runtime::String&, 
const tvm::runtime::Map<tvm::te::Tensor, tvm::tir::Buffer>&, bool)> >
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1671
   [2022-11-02T15:27:34.310Z]   19: run<>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1631
   [2022-11-02T15:27:34.310Z]   18: 
run<tvm::runtime::TVMMovableArgValueWithContext_>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1631
   [2022-11-02T15:27:34.310Z]   17: 
run<tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1631
   [2022-11-02T15:27:34.310Z]   16: 
run<tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1631
   [2022-11-02T15:27:34.310Z]   15: 
run<tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1631
   [2022-11-02T15:27:34.310Z]   14: 
run<tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_, 
tvm::runtime::TVMMovableArgValueWithContext_>
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1646
   [2022-11-02T15:27:34.310Z]   13: operator()
   [2022-11-02T15:27:34.310Z]         at ../src/driver/driver_api.cc:391
   [2022-11-02T15:27:34.310Z]   12: tvm::LowerSchedule(tvm::te::Schedule, 
tvm::runtime::Array<tvm::runtime::ObjectRef, void> const&, 
std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > 
const&, std::unordered_map<tvm::te::Tensor, tvm::tir::Buffer, 
std::hash<tvm::te::Tensor>, std::equal_to<tvm::te::Tensor>, 
std::allocator<std::pair<tvm::te::Tensor const, tvm::tir::Buffer> > > const&, 
tvm::GlobalVarSupply, bool)
   [2022-11-02T15:27:34.310Z]         at ../src/driver/driver_api.cc:377
   [2022-11-02T15:27:34.310Z]   11: tvm::LowerWithPassList(tvm::IRModule, 
tvm::runtime::Array<tvm::transform::Pass, void>)
   [2022-11-02T15:27:34.310Z]         at ../src/driver/driver_api.cc:272
   [2022-11-02T15:27:34.310Z]   10: 
tvm::transform::Pass::operator()(tvm::IRModule) const
   [2022-11-02T15:27:34.310Z]         at ../src/ir/transform.cc:258
   [2022-11-02T15:27:34.310Z]   9: 
tvm::transform::Pass::operator()(tvm::IRModule, tvm::transform::PassContext 
const&) const
   [2022-11-02T15:27:34.310Z]         at ../src/ir/transform.cc:274
   [2022-11-02T15:27:34.310Z]   8: 
tvm::transform::SequentialNode::operator()(tvm::IRModule, 
tvm::transform::PassContext const&) const
   [2022-11-02T15:27:34.310Z]         at ../src/ir/transform.cc:453
   [2022-11-02T15:27:34.310Z]   7: 
tvm::transform::Pass::operator()(tvm::IRModule, tvm::transform::PassContext 
const&) const
   [2022-11-02T15:27:34.310Z]         at ../src/ir/transform.cc:274
   [2022-11-02T15:27:34.310Z]   6: 
tvm::tir::transform::PrimFuncPassNode::operator()(tvm::IRModule, 
tvm::transform::PassContext const&) const
   [2022-11-02T15:27:34.310Z]         at ../src/tir/ir/transform.cc:100
   [2022-11-02T15:27:34.310Z]   5: 
tvm::runtime::TypedPackedFunc<tvm::tir::PrimFunc (tvm::tir::PrimFunc, 
tvm::IRModule, tvm::transform::PassContext)>::operator()(tvm::tir::PrimFunc, 
tvm::IRModule, tvm::transform::PassContext) const
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1750
   [2022-11-02T15:27:34.310Z]   4: tvm::tir::PrimFunc 
tvm::runtime::detail::typed_packed_call_dispatcher<tvm::tir::PrimFunc>::run<tvm::tir::PrimFunc,
 tvm::IRModule, tvm::transform::PassContext>(tvm::runtime::PackedFunc const&, 
tvm::tir::PrimFunc&&, tvm::IRModule&&, tvm::transform::PassContext&&)
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1694
   [2022-11-02T15:27:34.310Z]   3: tvm::runtime::TVMRetValue 
tvm::runtime::PackedFunc::operator()<tvm::tir::PrimFunc, tvm::IRModule, 
tvm::transform::PassContext>(tvm::tir::PrimFunc&&, tvm::IRModule&&, 
tvm::transform::PassContext&&) const
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1618
   [2022-11-02T15:27:34.310Z]   2: 
tvm::runtime::PackedFuncObj::CallPacked(tvm::runtime::TVMArgs, 
tvm::runtime::TVMRetValue*) const
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1217
   [2022-11-02T15:27:34.310Z]   1: Call
   [2022-11-02T15:27:34.310Z]         at 
../include/tvm/runtime/packed_func.h:1213
   [2022-11-02T15:27:34.310Z]   0: operator()
   [2022-11-02T15:27:34.310Z]         at ../src/runtime/c_runtime_api.cc:534
   [2022-11-02T15:27:34.310Z]   File "tvm/_ffi/_cython/./packed_func.pxi", line 
56, in tvm._ffi._cy3.core.tvm_callback
   [2022-11-02T15:27:34.310Z]   File 
"/workspace/python/tvm/autotvm/measure/measure_methods.py", line 871, in 
verify_pass
   [2022-11-02T15:27:34.310Z]     raise InstantiationError("Skipped because of 
invalid gpu kernel")
   [2022-11-02T15:27:34.310Z] tvm.autotvm.task.space.InstantiationError: 
Skipped because of invalid gpu kernel  [('tile_f', [-1, 16, 32, 1]), ('tile_y', 
[-1, 1, 1, 1]), ('tile_x', [-1, 7, 1, 1]), ('tile_rc', [-1, 2, 64]), 
('tile_ry', [-1, 1, 1]), ('tile_rx', [-1, 1, 3]), ('auto_unroll_max_step', 0), 
('unroll_explicit', 0)],None,1324444
   [2022-11-02T15:27:34.310Z] WARNING:root:Could not find any valid schedule 
for task Task(func_name=tutorial/conv2d_no_batching, args=(1, 7, 7, 512, 512, 
3, 3, (1, 1), (1, 1)), kwargs={}, workload=('tutorial/conv2d_no_batching', 1, 
7, 7, 512, 512, 3, 3, (1, 1), (1, 1))). A file containing the errors has been 
written to /tmp/tvm_tuning_errors_0x9nxpws.log.
   [2022-11-02T15:27:34.310Z] DEBUG:autotvm:Finish loading 20 records
   [2022-11-02T15:27:34.310Z] WARNING:autotvm:Cannot find config for 
target=cuda -keys=cuda,gpu -arch=sm_75 -max_num_threads=1024 
-thread_warp_size=32, workload=('tutorial/conv2d_no_batching', 1, 7, 7, 512, 
512, 3, 3, (1, 1), (1, 1)). A fallback configuration is used, which may bring 
great performance regression.
   [2022-11-02T15:27:34.310Z] DEBUG:autotvm:Finish loading 20 records
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to