Cookiee235 opened a new issue, #17934:
URL: https://github.com/apache/tvm/issues/17934

   ### Actual behavior
   
   ```
   2025-05-09 18:24:12 [INFO] Logging directory: ./tune_tmp/logs
   2025-05-09 18:24:15 [INFO] LocalBuilder: max_workers = 32
   2025-05-09 18:24:16 [INFO] LocalRunner: max_workers = 1
   2025-05-09 18:24:16 [INFO] [task_scheduler.cc:159] Initializing Task #0: 
"main"
   2025-05-09 18:24:18 [INFO] [task_scheduler.cc:320]
    ID | Name |   FLOP | Weight | Speed (GFLOPS) | Latency (us) | Weighted 
Latency (us) | Trials | Done
   
-----------------------------------------------------------------------------------------------------
     0 | main | 434176 |      1 |            N/A |          N/A |               
    N/A |      0 |
   
-----------------------------------------------------------------------------------------------------
   Total trials: 0
   Total latency (us): 0
   
   2025-05-09 18:24:18 [INFO] [task_scheduler.cc:180] TaskScheduler picks Task 
#0: "main"
   Traceback (most recent call last):
     File 
"/data/qshenaf/remote_pc/TirFuzz/bugs/05-03_20-50/topi.nn.pool_grad_0.py", line 
11, in <module>
       database = ms.tir_integration.tune_tir(mod=sch.mod, target='llvm 
--num-cores=16', work_dir='./tune_tmp', max_trials_global=1, 
num_trials_per_iter=1)
                  
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File "/data/qshenaf/envs/tvm/python/tvm/meta_schedule/tir_integration.py", 
line 146, in tune_tir
       return tune_tasks(
              ^^^^^^^^^^^
     File "/data/qshenaf/envs/tvm/python/tvm/meta_schedule/tune.py", line 122, 
in tune_tasks
       task_scheduler.tune(
     File 
"/data/qshenaf/envs/tvm/python/tvm/meta_schedule/task_scheduler/task_scheduler.py",
 line 132, in tune
       _ffi_api.TaskSchedulerTune(  # type: ignore # pylint: disable=no-member
     File "tvm/_ffi/_cython/./packed_func.pxi", line 339, in 
tvm._ffi._cy3.core.PackedFuncBase.__call__
     File "tvm/_ffi/_cython/./packed_func.pxi", line 284, in 
tvm._ffi._cy3.core.FuncCall
     File "tvm/_ffi/_cython/./base.pxi", line 185, in 
tvm._ffi._cy3.core.CHECK_CALL
     File "/data/qshenaf/envs/tvm/python/tvm/_ffi/base.py", line 468, in 
raise_last_ffi_error
       raise py_err
     File 
"/data/qshenaf/envs/tvm/src/meta_schedule/task_scheduler/gradient_based.cc", 
line 54, in 
tvm::meta_schedule::GradientBasedNode::Tune(tvm::runtime::Array<tvm::meta_schedule::TuneContext,
 void>, tvm::runtime::Array<tvm::FloatImm, void>, int, int, int, 
tvm::meta_schedule::Builder, tvm::meta_schedule::Runner, 
tvm::runtime::Array<tvm::meta_schedule::MeasureCallback, void>, 
tvm::runtime::Optional<tvm::meta_schedule::Database>, 
tvm::runtime::Optional<tvm::meta_schedule::CostModel>)
       TaskSchedulerNode::Tune(tasks, task_weights, max_trials_global, 
max_trials_per_task,
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File 
"/data/qshenaf/envs/tvm/src/meta_schedule/task_scheduler/task_scheduler.cc", 
line 190, in 
tvm::meta_schedule::TaskSchedulerNode::Tune(tvm::runtime::Array<tvm::meta_schedule::TuneContext,
 void>, tvm::runtime::Array<tvm::FloatImm, void>, int, int, int, 
tvm::meta_schedule::Builder, tvm::meta_schedule::Runner, 
tvm::runtime::Array<tvm::meta_schedule::MeasureCallback, void>, 
tvm::runtime::Optional<tvm::meta_schedule::Database>, 
tvm::runtime::Optional<tvm::meta_schedule::CostModel>)
       task->ctx->search_strategy.value()->GenerateMeasureCandidates()) {
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File 
"/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc",
 line 447, in 
tvm::meta_schedule::EvolutionarySearchNode::GenerateMeasureCandidates()
       return this->state_->GenerateMeasureCandidates();
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File 
"/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc",
 line 717, in 
tvm::meta_schedule::EvolutionarySearchNode::State::GenerateMeasureCandidates()
       std::vector<Schedule> unmeasured = SampleInitPopulation(pop - 
measured.size());
                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File 
"/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc",
 line 524, in 
tvm::meta_schedule::EvolutionarySearchNode::State::SampleInitPopulation(int)
       support::parallel_for_dynamic(0, num, self->ctx_->num_threads, 
f_proc_unmeasured);
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
     File "/data/qshenaf/envs/tvm/src/support/parallel_for.cc", line 128, in 
tvm::support::parallel_for_dynamic(int, int, int, std::function<void (int, 
int)> const&)
       LOG(FATAL) << "RuntimeError: parallel_for_dynamic error with " << 
e.what();
                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^
   tvm._ffi.base.TVMError: Traceback (most recent call last):
     5: 
tvm::meta_schedule::GradientBasedNode::Tune(tvm::runtime::Array<tvm::meta_schedule::TuneContext,
 void>, tvm::runtime::Array<tvm::FloatImm, void>, int, int, int, 
tvm::meta_schedule::Builder, tvm::meta_schedule::Runner, 
tvm::runtime::Array<tvm::meta_schedule::MeasureCallback, void>, 
tvm::runtime::Optional<tvm::meta_schedule::Database>, 
tvm::runtime::Optional<tvm::meta_schedule::CostModel>)
           at 
/data/qshenaf/envs/tvm/src/meta_schedule/task_scheduler/gradient_based.cc:54
     4: 
tvm::meta_schedule::TaskSchedulerNode::Tune(tvm::runtime::Array<tvm::meta_schedule::TuneContext,
 void>, tvm::runtime::Array<tvm::FloatImm, void>, int, int, int, 
tvm::meta_schedule::Builder, tvm::meta_schedule::Runner, 
tvm::runtime::Array<tvm::meta_schedule::MeasureCallback, void>, 
tvm::runtime::Optional<tvm::meta_schedule::Database>, 
tvm::runtime::Optional<tvm::meta_schedule::CostModel>)
           at 
/data/qshenaf/envs/tvm/src/meta_schedule/task_scheduler/task_scheduler.cc:190
     3: tvm::meta_schedule::EvolutionarySearchNode::GenerateMeasureCandidates()
           at 
/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc:447
     2: 
tvm::meta_schedule::EvolutionarySearchNode::State::GenerateMeasureCandidates()
           at 
/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc:717
     1: 
tvm::meta_schedule::EvolutionarySearchNode::State::SampleInitPopulation(int)
           at 
/data/qshenaf/envs/tvm/src/meta_schedule/search_strategy/evolutionary_search.cc:524
     0: tvm::support::parallel_for_dynamic(int, int, int, std::function<void 
(int, int)> const&)
           at /data/qshenaf/envs/tvm/src/support/parallel_for.cc:128
     File "/data/qshenaf/envs/tvm/src/support/parallel_for.cc", line 128
   RuntimeError: parallel_for_dynamic error with ScheduleError: (not rendered)
   ```
   
   ### Environment
   
   tvm-0.21.dev0'
   
   ### Steps to reproduce
   
   ```
   import tvm
   from tvm import te, topi, tir
   from tvm import meta_schedule as ms
   
   grads = te.placeholder((1, 16, 32, 32), dtype='float32', name='grads')
   data = te.placeholder((1, 16, 32, 32), dtype='float32', name='data')
   op_config = {'grads': grads, 'data': data, 'kernel': [3, 3], 'stride': [2, 
2], 'padding': [1, 1, 1, 1], 'pool_type': 'max', 'ceil_mode': False, 
'count_include_pad': False, 'layout': 'NCHW', }
   op_output = topi.nn.pool_grad(**op_config)
   sch = tir.Schedule(te.create_prim_func([grads, data, 
op_output]).with_attr('target', tvm.target.Target('llvm')))
   
   database = ms.tir_integration.tune_tir(mod=sch.mod, target='llvm 
--num-cores=16', work_dir='./tune_tmp', max_trials_global=1, 
num_trials_per_iter=1)
   sch = ms.tir_integration.compile_tir(database, sch.mod, 'llvm 
--num-cores=16')
   
   ```
   ### Triage
   
   * needs-triage
   * tune:meta_schedule
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to