Wheest opened a new issue, #11233:
URL: https://github.com/apache/tvm/issues/11233
### Expected behavior
Compiling a model with `debug_executor` allows one to run the model and get
the tracing output.
### Actual behavior
Process fails during the module creation:
```
Traceback (most recent call last):
File "tvm_profiler_simple.py", line 98, in <module>
main(args)
File "tvm_profiler_simple.py", line 64, in main
m = debug_executor.create(lib.graph_json, lib, dev,
dump_root="/tmp/tvmdbg")
File "/app/source/tvm/python/tvm/contrib/debugger/debug_executor.py", line
70, in create
func_obj = fcreate(graph_json_str, libmod, *device_type_id)
File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 223,
in __call__
values, tcodes, num_args = _make_tvm_args(args, temp_args)
File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 188,
in _make_tvm_args
raise TypeError("Don't know how to handle type %s" % type(arg))
TypeError: Don't know how to handle type <class
'tvm.relay.backend.executor_factory.GraphExecutorFactoryModule'>
```
### Environment
x86 platform, TVM v0.8.
### Steps to reproduce
[This
script](https://gist.github.com/Wheest/9ad2d6a47bbd2cfaa4be530c68ba2f6c) shows
the issue.
The gist has three modes: `["tutorial", "alt", "normal"]`, invoked with
`python tvm_profiler_simple.py --mode tutorial`.
- `normal` inference works of course
- `tutorial` is the approach in the current documentation, with fails with
output [1]
- `alt` is the approach used in [this docs
PR](https://github.com/apache/tvm/pull/11231) which works, with example output
[2]
As the [discussion in the forum
says](https://discuss.tvm.apache.org/t/runnig-a-model-with-tvm-debugger/9869/9?u=wheest),
there does not appear to be a canonical way of creating the debugger, but the
approach in the docs should at least work, until there has been a refactor to
produce an unambiguous canonical approach.
[1] `tutorial` sample output:
```
Traceback (most recent call last):
File "tvm_profiler_simple.py", line 98, in <module>
main(args)
File "tvm_profiler_simple.py", line 64, in main
m = debug_executor.create(lib.graph_json, lib, dev,
dump_root="/tmp/tvmdbg")
File "/app/source/tvm/python/tvm/contrib/debugger/debug_executor.py", line
70, in create
func_obj = fcreate(graph_json_str, libmod, *device_type_id)
File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 223,
in __call__
values, tcodes, num_args = _make_tvm_args(args, temp_args)
File "/app/source/tvm/python/tvm/_ffi/_ctypes/packed_func.py", line 188,
in _make_tvm_args
raise TypeError("Don't know how to handle type %s" % type(arg))
TypeError: Don't know how to handle type <class
'tvm.relay.backend.executor_factory.GraphExecutorFactoryModule'>
```
[2] `alt` sample output
```
[19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:103:
Iteration: 0
[19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:108:
Op #0 tvmgen_default_fused_layout_transform: 29.3538 us/iter
[19:53:25] ../src/runtime/graph_executor/debug/graph_executor_debug.cc:108:
Op #1 tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu: 3009 us/iter
...
Node Name Ops
Time(us) Time(%) Shape
Inputs Outputs
--------- ---
-------- ------- -----
------ -------
tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu
tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_nn_relu 3009.0
7.115 (1, 2, 112, 112, 32) 3 1
tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_add_nn_relu_4
tvmgen_default_fused_nn_contrib_conv2d_NCHWc_add_add_nn_relu_4 2835.26
6.704 (1, 16, 14, 14, 16) 4 1
...
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]