I tried for all available queues and get the same messages.
$ qsub -w v c3-parallel-pgimpich2-demo.sh
Unable to run job: Job 144914 cannot run in queue "cc001-t001" because
it is not contained in its hard queue list (-q)
Job 144914 cannot run in queue "q0-em64t-ge" because it is not contained
in its hard queue list (-q)
Job 144914 cannot run in queue "q0-em64t-ib" because it is not contained
in its hard queue list (-q)
Job 144914 cannot run in PE "mpich" because it only offers 0 slots
verification: no suitable queues.
Exiting.
$ qsub -q q0-em64t-ib -w v c3-parallel-pgimpich2-demo.sh
Unable to run job: Job 144915 cannot run in queue "cc001-t001" because
it is not contained in its hard queue list (-q)
Job 144915 cannot run in queue "q0-em64t-ddr" because it is not
contained in its hard queue list (-q)
Job 144915 cannot run in queue "q0-em64t-ge" because it is not contained
in its hard queue list (-q)
Job 144915 cannot run in PE "mpich" because it only offers 0 slots
verification: no suitable queues.
Exiting.
$ qsub -q q0-em64t-ge -w v c3-parallel-pgimpich2-demo.sh
Unable to run job: Job 144916 cannot run in queue "cc001-t001" because
it is not contained in its hard queue list (-q)
Job 144916 cannot run in queue "q0-em64t-ddr" because it is not
contained in its hard queue list (-q)
Job 144916 cannot run in queue "q0-em64t-ib" because it is not contained
in its hard queue list (-q)
Job 144916 cannot run in PE "mpich" because it only offers 0 slots
verification: no suitable queues.
Exiting.
I created a temporary PE and used qalter to run those queued jobs.
Best Regards,
Lu
於 2011/12/7 下午 07:32, Reuti 提到:
Hi,
Am 07.12.2011 um 08:13 schrieb wzlu:
The same problem occur again.
I try command "qsub -w v" and get following message.
Unable to run job: Job 144878 cannot run in queue "cc001-t001" because it is
not contained in its hard queue list (-q)
Job 144878 cannot run in queue "q0-em64t-ge" because it is not contained in its
hard queue list (-q)
Job 144878 cannot run in queue "q0-em64t-ib" because it is not contained in its
hard queue list (-q)
Job 144878 cannot run in PE "mpich" because it only offers 0 slots
so, the PE you requested is "mpich". Did you request any queue in the `qsub`
command, and is the PE attached to this queue?
-- Reuti
verification: no suitable queues.
Exiting.
Have any idea? Thanks.
Best Regards,
Lu
於 2011/10/7 下午 09:04, Reuti 提到:
Am 06.10.2011 um 14:40 schrieb Jesse Becker:
I ran into this a few months ago, and it had almost nothing to do with
PE slots. Unfortunately, I can't recall what I did to fix it either.
Try submitting test jobs with "-w v" and "-w p" to get more of an idea
of what's going on.
Yes, this needs to be investigated by hand. It's an RFE to get a better
scheduling output. Like here, you would like to know why the slots couldn't be
allocated. That there are only zero slots avilable, is the result of another
limit already.
Could be memory, RQS, slots, ...
-- Reuti
On Thu, Oct 06, 2011 at 04:39:39AM -0400, wzlu wrote:
Dear All,
There are 144 nodes in my queue and I configured 1 slot for each node. That is
144 nodes with 144 slots.
The PE is used 121 slots now. One job need 12 PE's slots and there are enough
nodes and slots for this job.
But it queued by "cannot run in PE "mpich" because it only offers 0 slots".
Configure as following:
$ qconf -sp mpich
pe_name mpich
slots 81920
user_lists NONE
xuser_lists NONE
start_proc_args /bin/true
stop_proc_args /bin/true
allocation_rule $round_robin
control_slaves TRUE
job_is_first_task FALSE
urgency_slots min
$ qconf -ssconf
algorithm default
schedule_interval 0:0:5
maxujobs 0
queue_sort_method load
job_load_adjustments NONE
load_adjustment_decay_time 0:7:30
load_formula slots
schedd_job_info true
flush_submit_sec 0
flush_finish_sec 0
params none
reprioritize_interval 0:0:0
halftime 168
usage_weight_list cpu=1.000000,mem=0.000000,io=0.000000
compensation_factor 5.000000
weight_user 0.250000
weight_project 0.250000
weight_department 0.250000
weight_job 0.250000
weight_tickets_functional 0
weight_tickets_share 0
share_override_tickets TRUE
share_functional_shares TRUE
max_functional_jobs_to_schedule 200
report_pjob_tickets TRUE
max_pending_tasks_per_job 50
halflife_decay_list none
policy_hierarchy OFS
weight_ticket 0.010000
weight_waiting_time 0.000000
weight_deadline 3600000.000000
weight_urgency 0.100000
weight_priority 1.000000
max_reservation 0
default_duration 00:15:00
How to fix this problem. Thanks a lot.
Best Regards,
Lu
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
--
Jesse Becker
NHGRI Linux support (Digicon Contractor)
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users