[TVM Discuss] [Questions] TVM is 10x slower than simple Mxnet model

2020-07-01 Thread Yao Wang via TVM Discuss
Multi-core should significantly improve performance. Can you use htop to watch the cpu usage? --- [Visit Topic](https://discuss.tvm.ai/t/tvm-is-10x-slower-than-simple-mxnet-model/7154/6) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these

[TVM Discuss] [Questions] TVM is 10x slower than simple Mxnet model

2020-07-01 Thread Yao Wang via TVM Discuss
How many CPU cores are you using? Latency number looks like single core inference. --- [Visit Topic](https://discuss.tvm.ai/t/tvm-is-10x-slower-than-simple-mxnet-model/7154/4) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails,

[TVM Discuss] [Questions] TVM is 10x slower than simple Mxnet model

2020-07-01 Thread Yao Wang via TVM Discuss
Change target to "llvm -mcpu=skylake-avx512"? --- [Visit Topic](https://discuss.tvm.ai/t/tvm-is-10x-slower-than-simple-mxnet-model/7154/2) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] Graph tuning a TE computation? te.op to relay.Expr?

2020-06-25 Thread Yao Wang via TVM Discuss
https://tvm.apache.org/docs/dev/relay_add_op.html --- [Visit Topic](https://discuss.tvm.ai/t/graph-tuning-a-te-computation-te-op-to-relay-expr/7037/5) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] has no attribute name_hint when importing SSD-ResNet34 from TensorFlow

2020-06-17 Thread Yao Wang via TVM Discuss
@cu Would you mind file a patch for this? --- [Visit Topic](https://discuss.tvm.ai/t/class-tvm-relay-expr-call-has-no-attribute-name-hint-when-importing-ssd-resnet34-from-tensorflow/6989/6) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from

[TVM Discuss] [Questions] has no attribute name_hint when importing SSD-ResNet34 from TensorFlow

2020-06-17 Thread Yao Wang via TVM Discuss
This means padlist is actually symbolic. You can try something like ```padlist = _infer_value(inputs[1], params, mod)``` to see whether it can be pre-computed. --- [Visit

[TVM Discuss] [Questions] has no attribute name_hint when importing SSD-ResNet34 from TensorFlow

2020-06-16 Thread Yao Wang via TVM Discuss
Can you print full trace? --- [Visit Topic](https://discuss.tvm.ai/t/class-tvm-relay-expr-call-has-no-attribute-name-hint-when-importing-ssd-resnet34-from-tensorflow/6989/2) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails,

[TVM Discuss] [Questions] [BUG] TVM Generate wrong OpenCL code

2020-06-01 Thread Yao Wang via TVM Discuss
https://github.com/apache/incubator-tvm/issues/5710 --- [Visit Topic](https://discuss.tvm.ai/t/bug-tvm-generate-wrong-opencl-code/6658/5) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] How to setting model compiled from pytorch with mutable input size

2020-05-28 Thread Yao Wang via TVM Discuss
Dynamic input shape will involve dynamic shape kernel codegen, which is still WIP. --- [Visit Topic](https://discuss.tvm.ai/t/how-to-setting-model-compiled-from-pytorch-with-mutable-input-size/6827/4) to respond. You are receiving this because you enabled mailing list mode. To

[TVM Discuss] [Questions] [AutoTVM]Tuning fails for an ONNX network on x86 CPU in tune_graph KeyError: ‘topi_op’

2020-05-27 Thread Yao Wang via TVM Discuss
It looks like ```multiply``` is not recognized as mutiple input op. Can you dig into https://github.com/apache/incubator-tvm/blob/master/python/tvm/autotvm/graph_tuner/base_graph_tuner.py#L167 to see why it is not recognized? --- [Visit

[TVM Discuss] [Questions] [Question] After auto tune, the speed of resnest101 and resnet101 have big difference

2020-05-21 Thread Yao Wang via TVM Discuss
What is your target device? --- [Visit Topic](https://discuss.tvm.ai/t/question-after-auto-tune-the-speed-of-resnest101-and-resnet101-have-big-difference/6748/2) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] Issue with static tensor array

2020-05-18 Thread Yao Wang via TVM Discuss
I'm refactoring tf frontend tensor array and will fix these issues. --- [Visit Topic](https://discuss.tvm.ai/t/issue-with-static-tensor-array/6333/12) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] Questions regarding the ATC19 paper "Optimizing CNN Model Inference on CPUs"

2020-04-22 Thread Yao Wang via TVM Discuss
Integrating MKLDNN into deep learning framework requires special handle for the fusion of those operators which will be accelerated using MKLDNN kernel, sometimes even patten matching nn blocks. It's quite difficult to develop a general fusion rule in this situation. --- [Visit

[TVM Discuss] [Questions] Questions regarding the ATC19 paper "Optimizing CNN Model Inference on CPUs"

2020-04-22 Thread Yao Wang via TVM Discuss
Graph tuning is to balance the tradeoff between data layout overhead and kernel with fast schedule, MKLDNN needs to be integrated with deep learning framework to execute a NN. In this process, overheads such as data layout transformation can generate. Another major overhead comes from

[TVM Discuss] [Questions] Questions regarding the ATC19 paper "Optimizing CNN Model Inference on CPUs"

2020-04-22 Thread Yao Wang via TVM Discuss
@moderato The performance advantage comes from the joint optimization of both kernel and graph level optimization. For conv2d kernels, in lots of cases MKLDNN outperforms TVM. However, with graph tuning we can achieve better e2e performance by carefully arranging data layout. --- [Visit

[TVM Discuss] [Questions] Can TVM now support batched inference? Autotvm runs twice as long as tensorflow

2020-04-20 Thread Yao Wang via TVM Discuss
You don't need graph tuning while using cblas. --- [Visit Topic](https://discuss.tvm.ai/t/can-tvm-now-support-batched-inference-autotvm-runs-twice-as-long-as-tensorflow/6405/8) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails,

[TVM Discuss] [Questions] Issue with static tensor array

2020-04-17 Thread Yao Wang via TVM Discuss
I'm fixing some issues regarding to tf ssd models and will submit a PR soon. --- [Visit Topic](https://discuss.tvm.ai/t/issue-with-static-tensor-array/6333/6) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click

[TVM Discuss] [Questions] Issue with static tensor array

2020-04-11 Thread Yao Wang via TVM Discuss
Can you provide the full trace? Also it should be related to certain tensor array op. It would be better if you can provide a minimal case. --- [Visit Topic](https://discuss.tvm.ai/t/issue-with-static-tensor-array/6333/2) to respond. You are receiving this because you enabled mailing

[TVM Discuss] [Questions] [autoTVM][Graph tuner] Running graph tuner without autoTVM

2020-04-08 Thread Yao Wang via TVM Discuss
Currently log files in tophub just store the best schedule for each workload. The idea for graph tuner is to select a schedule from topk(usually 20 - 30) best schedules from a workload so that we can minimize data layout transformation overhead. Thus we want to first do autotuning. ---

[TVM Discuss] [Questions] Disabling LLVM Unrolling

2020-03-22 Thread Yao Wang via TVM Discuss
This is a quite valuable topic which can help us figure out what kind of information related to optimization we can get from TVM IR itself, after all the LLVM optimization pass applied. For x86 conv2d, my observation is that the work llvm unrolling is doing can be implemented in TVM schedule

[TVM Discuss] [Questions] Calculate number of FLOP error

2019-10-11 Thread Yao Wang via TVM Discuss
Did you call cfg.add_flop to add flops number? --- [Visit Topic](http://tracking.discuss.tvm.ai/tracking/click?d=2zHJ-oLuPzy-Sp0s1DTQwn9qeKjUgBbdu3W75xbqGy881gxEFiNQ5hRjAKkz5tGT9Ou4R8dj6zHjz4rG9XUgXoJYT0a7H5b0HfcHZnxJbQFbREDwZmkJ7PDOWLim8tqzaQFhKHtut8CeeAGHTonhFNJZLxxYTwUdiJl-J8aLaq9G0)

[TVM Discuss] [Questions] How to map lowered Function with Task Name in AutoTVM?

2019-08-12 Thread Yao Wang via TVM Discuss
You can print the workload of each conv2d: tsk.workload. --- [Visit Topic](https://discuss.tvm.ai/t/how-to-map-lowered-function-with-task-name-in-autotvm/3704/2) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click