Hi,

Am 07.12.2011 um 08:13 schrieb wzlu:

> The same problem occur again.
> 
> I try command "qsub -w v" and get following message.
> Unable to run job: Job 144878 cannot run in queue "cc001-t001" because it is 
> not contained in its hard queue list (-q)
> Job 144878 cannot run in queue "q0-em64t-ge" because it is not contained in 
> its hard queue list (-q)
> Job 144878 cannot run in queue "q0-em64t-ib" because it is not contained in 
> its hard queue list (-q)
> Job 144878 cannot run in PE "mpich" because it only offers 0 slots

so, the PE you requested is "mpich". Did you request any queue in the `qsub` 
command, and is the PE attached to this queue?

-- Reuti


> verification: no suitable queues.
> Exiting.
> 
> Have any idea? Thanks.
> 
> Best Regards,
> Lu
> 
> 於 2011/10/7 下午 09:04, Reuti 提到:
>> Am 06.10.2011 um 14:40 schrieb Jesse Becker:
>> 
>>> I ran into this a few months ago, and it had almost nothing to do with
>>> PE slots. Unfortunately, I can't recall what I did to fix it either.
>>> Try submitting test jobs with "-w v" and "-w p" to get more of an idea
>>> of what's going on.
>> 
>> Yes, this needs to be investigated by hand. It's an RFE to get a better 
>> scheduling output. Like here, you would like to know why the slots couldn't 
>> be allocated. That there are only zero slots avilable, is the result of 
>> another limit already.
>> 
>> Could be memory, RQS, slots, ...
>> 
>> -- Reuti
>> 
>>> On Thu, Oct 06, 2011 at 04:39:39AM -0400, wzlu wrote:
>>>> Dear All,
>>>> 
>>>> There are 144 nodes in my queue and I configured 1 slot for each node. 
>>>> That is 144 nodes with 144 slots.
>>>> The PE is used 121 slots now. One job need 12 PE's slots and there are 
>>>> enough nodes and slots for this job.
>>>> But it queued by "cannot run in PE "mpich" because it only offers 0 slots".
>>>> 
>>>> Configure as following:
>>>> 
>>>> $ qconf -sp mpich
>>>> pe_name mpich
>>>> slots 81920
>>>> user_lists NONE
>>>> xuser_lists NONE
>>>> start_proc_args /bin/true
>>>> stop_proc_args /bin/true
>>>> allocation_rule $round_robin
>>>> control_slaves TRUE
>>>> job_is_first_task FALSE
>>>> urgency_slots min
>>>> 
>>>> $ qconf -ssconf
>>>> algorithm default
>>>> schedule_interval 0:0:5
>>>> maxujobs 0
>>>> queue_sort_method load
>>>> job_load_adjustments NONE
>>>> load_adjustment_decay_time 0:7:30
>>>> load_formula slots
>>>> schedd_job_info true
>>>> flush_submit_sec 0
>>>> flush_finish_sec 0
>>>> params none
>>>> reprioritize_interval 0:0:0
>>>> halftime 168
>>>> usage_weight_list cpu=1.000000,mem=0.000000,io=0.000000
>>>> compensation_factor 5.000000
>>>> weight_user 0.250000
>>>> weight_project 0.250000
>>>> weight_department 0.250000
>>>> weight_job 0.250000
>>>> weight_tickets_functional 0
>>>> weight_tickets_share 0
>>>> share_override_tickets TRUE
>>>> share_functional_shares TRUE
>>>> max_functional_jobs_to_schedule 200
>>>> report_pjob_tickets TRUE
>>>> max_pending_tasks_per_job 50
>>>> halflife_decay_list none
>>>> policy_hierarchy OFS
>>>> weight_ticket 0.010000
>>>> weight_waiting_time 0.000000
>>>> weight_deadline 3600000.000000
>>>> weight_urgency 0.100000
>>>> weight_priority 1.000000
>>>> max_reservation 0
>>>> default_duration 00:15:00
>>>> 
>>>> How to fix this problem. Thanks a lot.
>>>> 
>>>> Best Regards,
>>>> Lu
>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> [email protected]
>>>> https://gridengine.org/mailman/listinfo/users
>>> 
>>> 
>>> -- 
>>> Jesse Becker
>>> NHGRI Linux support (Digicon Contractor)
>>> _______________________________________________
>>> users mailing list
>>> [email protected]
>>> https://gridengine.org/mailman/listinfo/users
>> 
> 
> _______________________________________________
> users mailing list
> [email protected]
> https://gridengine.org/mailman/listinfo/users
> 


_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users

Reply via email to