Hello Michael

I have tried this as well with the same result.

I have even set properties for the quad core nodes and specified them in 
the pbs script to only select those nodes, but still the same problem.

I have created a new queue adding the properties of the quad core nodes 
to resources_max.nodes = 228:ppn=4:i965, however still the same result.

Thanks 
  
Nilesh Mistry 
Academic Computing Services 
[EMAIL PROTECTED] & TEL Campus 
Seneca College Of Applies Arts & Technology 
70 The Pond Road 
Toronto, Ontario 
M3J 3M6 Canada 
Phone 416 491 5050 ext 3788 
Fax 416 661 4695
http://acs.senecac.on.ca



Michael Edwards wrote:
> set queue workq resources_max.ncpus = 200
> set queue workq resources_max.nodect = 64
> set queue workq resources_max.nodes = 200:ppn=4
>
> This should probably be 50*4 + 14*2 = 228
>
> set queue workq resources_max.ncpus = 228
> set queue workq resources_max.nodect = 64
> set queue workq resources_max.nodes = 228:ppn=4
>
> Though you might want to try making two queues, I don't know how well
> torque deals with having different numbers of ppn on different nodes.
>
> On 9/18/07, Nilesh Mistry <[EMAIL PROTECTED]> wrote:
>   
>> I the following error after the using qsub
>>
>> qsub: Job exceeds queue resource limits
>>
>> If I change #PBS -l nodes=64 to #PBS -l nodes=60 i get the job submitted
>> and running and the if fails
>>
>> ################  qstat -f ############################
>>
>> Job Id: 924.master.atar.senecac.on.ca
>>     Job_Name = scaling_test
>>     Job_Owner = [EMAIL PROTECTED]
>>     job_state = R
>>     queue = parallel
>>     server = master.atar.senecac.on.ca
>>     Checkpoint = u
>>     ctime = Tue Sep 18 09:09:45 2007
>>     Error_Path =
>> master:/home/faculty/nilesh.mistry/pbs/multi/scaling_test/scal
>>         ing_test.err
>>     exec_host =
>> atarnode59.atar.senecac.on.ca/1+atarnode59.atar.senecac.on.ca/0
>>
>> +atarnode57.atar.senecac.on.ca/1+atarnode57.atar.senecac.on.ca/0+atarno
>>
>> de56.atar.senecac.on.ca/1+atarnode56.atar.senecac.on.ca/0+atarnode55.at
>>
>> ar.senecac.on.ca/1+atarnode55.atar.senecac.on.ca/0+atarnode54.atar.sene
>>
>> cac.on.ca/1+atarnode54.atar.senecac.on.ca/0+atarnode53.atar.senecac.on.
>>
>> ca/1+atarnode53.atar.senecac.on.ca/0+atarnode52.atar.senecac.on.ca/1+at
>>
>> arnode52.atar.senecac.on.ca/0+atarnode51.atar.senecac.on.ca/1+atarnode5
>>
>> 1.atar.senecac.on.ca/0+atarnode50.atar.senecac.on.ca/2+atarnode50.atar.
>>
>> senecac.on.ca/1+atarnode50.atar.senecac.on.ca/0+atarnode49.atar.senecac
>>
>> .on.ca/2+atarnode49.atar.senecac.on.ca/1+atarnode49.atar.senecac.on.ca/
>>
>> 0+atarnode48.atar.senecac.on.ca/2+atarnode48.atar.senecac.on.ca/1+atarn
>>
>> ode48.atar.senecac.on.ca/0+atarnode47.atar.senecac.on.ca/2+atarnode47.a
>>
>> tar.senecac.on.ca/1+atarnode47.atar.senecac.on.ca/0+atarnode45.atar.sen
>>
>> ecac.on.ca/2+atarnode45.atar.senecac.on.ca/1+atarnode45.atar.senecac.on
>>
>> .ca/0+atarnode44.atar.senecac.on.ca/2+atarnode44.atar.senecac.on.ca/1+a
>>
>> tarnode44.atar.senecac.on.ca/0+atarnode42.atar.senecac.on.ca/2+atarnode
>>
>> 42.atar.senecac.on.ca/1+atarnode42.atar.senecac.on.ca/0+atarnode41.atar
>>
>> .senecac.on.ca/2+atarnode41.atar.senecac.on.ca/1+atarnode41.atar.seneca
>>
>> c.on.ca/0+atarnode40.atar.senecac.on.ca/2+atarnode40.atar.senecac.on.ca
>>
>> /1+atarnode40.atar.senecac.on.ca/0+atarnode39.atar.senecac.on.ca/2+atar
>>
>> node39.atar.senecac.on.ca/1+atarnode39.atar.senecac.on.ca/0+atarnode38.
>>
>> atar.senecac.on.ca/2+atarnode38.atar.senecac.on.ca/1+atarnode38.atar.se
>>
>> necac.on.ca/0+atarnode37.atar.senecac.on.ca/2+atarnode37.atar.senecac.o
>>
>> n.ca/1+atarnode37.atar.senecac.on.ca/0+atarnode36.atar.senecac.on.ca/2+
>>
>> atarnode36.atar.senecac.on.ca/1+atarnode36.atar.senecac.on.ca/0+atarnod
>>
>> e35.atar.senecac.on.ca/2+atarnode35.atar.senecac.on.ca/1+atarnode35.ata
>>
>> r.senecac.on.ca/0+atarnode34.atar.senecac.on.ca/1+atarnode34.atar.senec
>>         ac.on.ca/0
>>     Hold_Types = n
>>     Join_Path = oe
>>     Keep_Files = n
>>     Mail_Points = abe
>>     Mail_Users = nilesh.mistry
>>     mtime = Tue Sep 18 09:09:46 2007
>>     Output_Path =
>> master:/home/faculty/nilesh.mistry/pbs/multi/scaling_test/sca
>>         ling_test.log
>>     Priority = 0
>>     qtime = Tue Sep 18 09:09:45 2007
>>     Rerunable = True
>>     Resource_List.cput = 10000:00:00
>>     Resource_List.mem = 64000mb
>>     Resource_List.ncpus = 1
>>     Resource_List.nodect = 60
>>     Resource_List.nodes = 60
>>     Resource_List.walltime = 10000:00:00
>>     Variable_List = PBS_O_HOME=/home/faculty/nilesh.mistry,
>>         PBS_O_LANG=en_CA.UTF-8,PBS_O_LOGNAME=nilesh.mistry,
>>
>> PBS_O_PATH=/usr/kerberos/bin:/opt/lam-7.1.2/bin:/usr/local/bin:/bin:/u
>>
>> sr/bin:/usr/X11R6/bin:/opt/env-switcher/bin:/opt/pvm3/lib:/opt/pvm3/lib
>>
>> /LINUX:/opt/pvm3/bin/LINUX:/opt/pbs/bin:/opt/pbs/lib/xpbs/bin:/opt/c3-4
>>
>> /:/home/faculty/nilesh.mistry/bin:/opt/maui/bin:/usr/lib/news/bin:/home
>>         /faculty/nilesh.mistry/scripts,
>>         PBS_O_MAIL=/var/spool/mail/nilesh.mistry,PBS_O_SHELL=/bin/bash,
>>         PBS_O_HOST=master.atar.senecac.on.ca,
>>         PBS_O_WORKDIR=/home/faculty/nilesh.mistry/pbs/multi/scaling_test,
>>         PBS_O_QUEUE=parallel
>>     etime = Tue Sep 18 09:09:45 2007
>>
>> ###################### Log file ##############################
>>
>> ------------------------------------------------------
>>  This job is allocated on 60 cpu(s)
>> Job is running on node(s):
>> atarnode59.atar.senecac.on.ca
>> atarnode59.atar.senecac.on.ca
>> atarnode57.atar.senecac.on.ca
>> atarnode57.atar.senecac.on.ca
>> atarnode56.atar.senecac.on.ca
>> atarnode56.atar.senecac.on.ca
>> atarnode55.atar.senecac.on.ca
>> atarnode55.atar.senecac.on.ca
>> atarnode54.atar.senecac.on.ca
>> atarnode54.atar.senecac.on.ca
>> atarnode53.atar.senecac.on.ca
>> atarnode53.atar.senecac.on.ca
>> atarnode52.atar.senecac.on.ca
>> atarnode52.atar.senecac.on.ca
>> atarnode51.atar.senecac.on.ca
>> atarnode51.atar.senecac.on.ca
>> atarnode50.atar.senecac.on.ca
>> atarnode50.atar.senecac.on.ca
>> atarnode50.atar.senecac.on.ca
>> atarnode49.atar.senecac.on.ca
>> atarnode49.atar.senecac.on.ca
>> atarnode49.atar.senecac.on.ca
>> atarnode48.atar.senecac.on.ca
>> atarnode48.atar.senecac.on.ca
>> atarnode48.atar.senecac.on.ca
>> atarnode47.atar.senecac.on.ca
>> atarnode47.atar.senecac.on.ca
>> atarnode47.atar.senecac.on.ca
>> atarnode45.atar.senecac.on.ca
>> atarnode45.atar.senecac.on.ca
>> atarnode45.atar.senecac.on.ca
>> atarnode44.atar.senecac.on.ca
>> atarnode44.atar.senecac.on.ca
>> atarnode44.atar.senecac.on.ca
>> atarnode42.atar.senecac.on.ca
>> atarnode42.atar.senecac.on.ca
>> atarnode42.atar.senecac.on.ca
>> atarnode41.atar.senecac.on.ca
>> atarnode41.atar.senecac.on.ca
>> atarnode41.atar.senecac.on.ca
>> atarnode40.atar.senecac.on.ca
>> atarnode40.atar.senecac.on.ca
>> atarnode40.atar.senecac.on.ca
>> atarnode39.atar.senecac.on.ca
>> atarnode39.atar.senecac.on.ca
>> atarnode39.atar.senecac.on.ca
>> atarnode38.atar.senecac.on.ca
>> atarnode38.atar.senecac.on.ca
>> atarnode38.atar.senecac.on.ca
>> atarnode37.atar.senecac.on.ca
>> atarnode37.atar.senecac.on.ca
>> atarnode37.atar.senecac.on.ca
>> atarnode36.atar.senecac.on.ca
>> atarnode36.atar.senecac.on.ca
>> atarnode36.atar.senecac.on.ca
>> atarnode35.atar.senecac.on.ca
>> atarnode35.atar.senecac.on.ca
>> atarnode35.atar.senecac.on.ca
>> atarnode34.atar.senecac.on.ca
>> atarnode34.atar.senecac.on.ca
>> PBS: qsub is running on master.atar.senecac.on.ca
>> PBS: originating queue is parallel
>> PBS: executing queue is parallel
>> PBS: working directory is /home/faculty/nilesh.mistry/pbs/multi/scaling_test
>> PBS: execution mode is PBS_BATCH
>> PBS: job identifier is 924.master.atar.senecac.on.ca
>> PBS: job name is scaling_test
>> PBS: node file is /var/spool/pbs/aux//924.master.atar.senecac.on.ca
>> PBS: current home directory is /home/faculty/nilesh.mistry
>> PBS: PATH =
>> /usr/kerberos/bin:/opt/lam-7.1.2/bin:/usr/local/bin:/bin:/usr/bin:/usr/X11R6/bin:/opt/env-switcher/bin:/opt/pvm3/lib:/opt/pvm3/lib/LINUX:/opt/pvm3/bi
>> n/LINUX:/opt/pbs/bin:/opt/pbs/lib/xpbs/bin:/opt/c3-4/:/home/faculty/nilesh.mistry/bin:/opt/maui/bin:/usr/lib/news/bin:/home/faculty/nilesh.mistry/scripts
>> ------------------------------------------------------
>> Mesh  1 of 60 is alive on atarnode59.atar.senecac.on.ca
>> Mesh 17 of 60 is alive on atarnode50.atar.senecac.on.ca
>> Mesh 18 of 60 is alive on atarnode50.atar.senecac.on.ca
>> Mesh  3 of 60 is alive on atarnode57.atar.senecac.on.ca
>> Mesh 50 of 60 is alive on atarnode37.atar.senecac.on.ca
>> Mesh 51 of 60 is alive on atarnode37.atar.senecac.on.ca
>> Mesh 58 of 60 is alive on atarnode35.atar.senecac.on.ca
>> Mesh 15 of 60 is alive on atarnode51.atar.senecac.on.ca
>> Mesh 56 of 60 is alive on atarnode35.atar.senecac.on.ca
>> Mesh 47 of 60 is alive on atarnode38.atar.senecac.on.ca
>> Mesh 41 of 60 is alive on atarnode40.atar.senecac.on.ca
>> Mesh 43 of 60 is alive on atarnode40.atar.senecac.on.ca
>> Mesh 23 of 60 is alive on atarnode48.atar.senecac.on.ca
>> Mesh 59 of 60 is alive on atarnode34.atar.senecac.on.ca
>> Mesh 44 of 60 is alive on atarnode39.atar.senecac.on.ca
>> Mesh 60 of 60 is alive on atarnode34.atar.senecac.on.ca
>> Mesh 26 of 60 is alive on atarnode47.atar.senecac.on.ca
>> Mesh 46 of 60 is alive on atarnode39.atar.senecac.on.ca
>> Mesh 42 of 60 is alive on atarnode40.atar.senecac.on.ca
>> Mesh 32 of 60 is alive on atarnode44.atar.senecac.on.ca
>> Mesh 20 of 60 is alive on atarnode49.atar.senecac.on.ca
>> Mesh 35 of 60 is alive on atarnode42.atar.senecac.on.ca
>> Mesh 53 of 60 is alive on atarnode36.atar.senecac.on.ca
>> Mesh 22 of 60 is alive on atarnode49.atar.senecac.on.ca
>> Mesh 19 of 60 is alive on atarnode50.atar.senecac.on.ca
>> Mesh 48 of 60 is alive on atarnode38.atar.senecac.on.ca
>> Mesh 37 of 60 is alive on atarnode42.atar.senecac.on.ca
>> Mesh 54 of 60 is alive on atarnode36.atar.senecac.on.ca
>> Mesh 55 of 60 is alive on atarnode36.atar.senecac.on.ca
>> Mesh 45 of 60 is alive on atarnode39.atar.senecac.on.ca
>> Mesh 29 of 60 is alive on atarnode45.atar.senecac.on.ca
>> Mesh 24 of 60 is alive on atarnode48.atar.senecac.on.ca
>> Mesh 30 of 60 is alive on atarnode45.atar.senecac.on.ca
>> Mesh 31 of 60 is alive on atarnode45.atar.senecac.on.ca
>> Mesh 52 of 60 is alive on atarnode37.atar.senecac.on.ca
>> Mesh 28 of 60 is alive on atarnode47.atar.senecac.on.ca
>> Mesh 36 of 60 is alive on atarnode42.atar.senecac.on.ca
>> Mesh 34 of 60 is alive on atarnode44.atar.senecac.on.ca
>> Mesh 38 of 60 is alive on atarnode41.atar.senecac.on.ca
>> Mesh 40 of 60 is alive on atarnode41.atar.senecac.on.ca
>> Mesh  5 of 60 is alive on atarnode56.atar.senecac.on.ca
>> Mesh 57 of 60 is alive on atarnode35.atar.senecac.on.ca
>> Mesh 13 of 60 is alive on atarnode52.atar.senecac.on.ca
>> Mesh  9 of 60 is alive on atarnode54.atar.senecac.on.ca
>> Mesh 39 of 60 is alive on atarnode41.atar.senecac.on.ca
>> Mesh  7 of 60 is alive on atarnode55.atar.senecac.on.ca
>> Mesh 10 of 60 is alive on atarnode54.atar.senecac.on.ca
>> Mesh  8 of 60 is alive on atarnode55.atar.senecac.on.ca
>> Mesh  4 of 60 is alive on atarnode57.atar.senecac.on.ca
>> Mesh  6 of 60 is alive on atarnode56.atar.senecac.on.ca
>> Mesh 11 of 60 is alive on atarnode53.atar.senecac.on.ca
>> Mesh 14 of 60 is alive on atarnode52.atar.senecac.on.ca
>> Mesh 12 of 60 is alive on atarnode53.atar.senecac.on.ca
>> Mesh 21 of 60 is alive on atarnode49.atar.senecac.on.ca
>> Mesh 16 of 60 is alive on atarnode51.atar.senecac.on.ca
>> Mesh 33 of 60 is alive on atarnode44.atar.senecac.on.ca
>> Mesh 49 of 60 is alive on atarnode38.atar.senecac.on.ca
>> Mesh 25 of 60 is alive on atarnode48.atar.senecac.on.ca
>> Mesh 27 of 60 is alive on atarnode47.atar.senecac.on.ca
>> Mesh  2 of 60 is alive on atarnode59.atar.senecac.on.ca
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>>
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>> ERROR: Number of meshes not equal to number of threads
>> ERROR: Number of meshes not equal to number of threads
>>
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>>
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>> ERROR: Number of meshes not equal to number of threads
>> ERROR: Number of meshes not equal to number of threads
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> ERROR: Number of meshes not equal to number of threads
>>
>> LAM 7.1.2/MPI 2 C++/ROMIO - Indiana University
>>
>>
>> Thanks
>>
>> Nilesh Mistry
>> Academic Computing Services
>> [EMAIL PROTECTED] & TEL Campus
>> Seneca College Of Applies Arts & Technology
>> 70 The Pond Road
>> Toronto, Ontario
>> M3J 3M6 Canada
>> Phone 416 491 5050 ext 3788
>> Fax 416 661 4695
>> http://acs.senecac.on.ca
>>
>>
>>
>> Michael Edwards wrote:
>>     
>>> What do you get when you do "qstat -f" on the job?  How many nodes is
>>> it actually getting?
>>>
>>> On 9/18/07, Nilesh Mistry <[EMAIL PROTECTED]> wrote:
>>>
>>>       
>>>> Micheal
>>>>
>>>> We have actually moved to a larger cluster of 64 nodes (50 quad core and
>>>> 14 dual opterons), there fore 220 processors available.  We are
>>>> submitting a job that requires 64 threads but still with the same
>>>> result.  Here are the files you requested.  I have already submitted to
>>>> torque user list.
>>>>
>>>> ####### PBS SCRIPT START#######
>>>>
>>>> #!/bin/sh -f
>>>> #PBS -l nodes=64
>>>> #PBS -N scaling_test
>>>> #PBS -e scaling_test.err
>>>> #PBS -o scaling_test.log
>>>> #PBS -j oe
>>>> #PBS -l mem=64000mb
>>>> #PBS -m abe
>>>> #PBS -q parallel
>>>>
>>>> NCPU=`wc -l < $PBS_NODEFILE`
>>>> echo ------------------------------------------------------
>>>> echo ' This job is allocated on '${NCPU}' cpu(s)'
>>>> echo 'Job is running on node(s): '
>>>> cat $PBS_NODEFILE
>>>> echo PBS: qsub is running on $PBS_O_HOST
>>>> echo PBS: originating queue is $PBS_O_QUEUE
>>>> echo PBS: executing queue is $PBS_QUEUE
>>>> echo PBS: working directory is $PBS_O_WORKDIR
>>>> echo PBS: execution mode is $PBS_ENVIRONMENT
>>>> echo PBS: job identifier is $PBS_JOBID
>>>> echo PBS: job name is $PBS_JOBNAME
>>>> echo PBS: node file is $PBS_NODEFILE
>>>> echo PBS: current home directory is $PBS_O_HOME
>>>> echo PBS: PATH = $PBS_O_PATH
>>>> echo ------------------------------------------------------
>>>> SERVER=$PBS_O_HOST
>>>> WORKDIR=$HOME/pbs/multi/scaling_test
>>>> cd ${WORKDIR}
>>>> cat $PBS_NODEFILE > nodes.list
>>>> lamboot -s -H $PBS_NODEFILE
>>>> mpirun -np $NCPU /opt/fds/fds5_mpi scaling_test.fds
>>>> lamhalt
>>>>
>>>> ####### PBS SCRIPT END #######
>>>>
>>>> ####### MAUI.CFG START #######
>>>> # maui.cfg 3.2.6p14
>>>>
>>>> SERVERHOST              master.atar.senecac.on.ca
>>>> # primary admin must be first in list
>>>> ADMIN1                  root
>>>> ADMIN3                  nilesh.mistry
>>>>
>>>>
>>>> # Resource Manager Definition
>>>>
>>>> RMCFG[master.atar.senecac.on.ca] TYPE=PBS
>>>>
>>>> # Allocation Manager Definition
>>>>
>>>> AMCFG[bank]  TYPE=NONE
>>>>
>>>> # full parameter docs at
>>>> http://clusterresources.com/mauidocs/a.fparameters.html
>>>> # use the 'schedctl -l' command to display current configuration
>>>>
>>>> RMPOLLINTERVAL  00:01:00
>>>>
>>>> SERVERPORT            42559
>>>> SERVERMODE            NORMAL
>>>>
>>>> # Admin: http://clusterresources.com/mauidocs/a.esecurity.html
>>>>
>>>>
>>>> LOGFILE               maui.log
>>>> LOGFILEMAXSIZE        10000000
>>>> LOGLEVEL              4
>>>> LOGFACILITY             fALL
>>>>
>>>> # Job Priority:
>>>> http://clusterresources.com/mauidocs/5.1jobprioritization.html
>>>>
>>>> QUEUETIMEWEIGHT       1
>>>>
>>>> # FairShare: http://clusterresources.com/mauidocs/6.3fairshare.html
>>>>
>>>> #FSPOLICY              PSDEDICATED
>>>> #FSDEPTH               7
>>>> #FSINTERVAL            86400
>>>> #FSDECAY               0.80
>>>>
>>>> # Throttling Policies:
>>>> http://clusterresources.com/mauidocs/6.2throttlingpolicies.html
>>>>
>>>> # NONE SPECIFIED
>>>>
>>>> # Backfill: http://clusterresources.com/mauidocs/8.2backfill.html
>>>>
>>>> BACKFILLPOLICY  ON
>>>> RESERVATIONPOLICY     CURRENTHIGHEST
>>>>
>>>> # the following are modified/added by Mehrdad 13 Sept 07
>>>> #NODEACCESSPOLICY       DEDICATED
>>>> NODEACCESSPOLICY        SHARED
>>>> JOBNODEMATCHPOLICY   EXACTPROC
>>>>
>>>> # Node Allocation:
>>>> http://clusterresources.com/mauidocs/5.2nodeallocation.html
>>>>
>>>> NODEALLOCATIONPOLICY  MINRESOURCE
>>>>
>>>> # QOS: http://clusterresources.com/mauidocs/7.3qos.html
>>>>
>>>> # QOSCFG[hi]  PRIORITY=100 XFTARGET=100 FLAGS=PREEMPTOR:IGNMAXJOB
>>>> # QOSCFG[low] PRIORITY=-1000 FLAGS=PREEMPTEE
>>>>
>>>> # Standing Reservations:
>>>> http://clusterresources.com/mauidocs/7.1.3standingreservations.html
>>>>
>>>> # SRSTARTTIME[test] 8:00:00
>>>> # SRENDTIME[test]   17:00:00
>>>> # SRDAYS[test]      MON TUE WED THU FRI
>>>> # SRTASKCOUNT[test] 20
>>>> # SRMAXTIME[test]   0:30:00
>>>>
>>>> # Creds: http://clusterresources.com/mauidocs/6.1fairnessoverview.html
>>>>
>>>> # USERCFG[DEFAULT]      FSTARGET=25.0
>>>> # USERCFG[john]         PRIORITY=100  FSTARGET=10.0-
>>>> # GROUPCFG[staff]       PRIORITY=1000 QLIST=hi:low QDEF=hi
>>>> # CLASSCFG[batch]       FLAGS=PREEMPTEE
>>>> # CLASSCFG[interactive] FLAGS=PREEMPTOR
>>>> USERCFG[DEFAULT]        MAXJOB=4
>>>> ####### MAUI.CFG  END #######
>>>>
>>>> ####### QMGR -c "PRINT SERVER MASTER" ########
>>>> #
>>>> # Create queues and set their attributes.
>>>> #
>>>> #
>>>> # Create and define queue serial
>>>> #
>>>> create queue serial
>>>> set queue serial queue_type = Execution
>>>> set queue serial resources_max.cput = 1000:00:00
>>>> set queue serial resources_max.mem = 3000mb
>>>> set queue serial resources_max.ncpus = 1
>>>> set queue serial resources_max.nodect = 1
>>>> set queue serial resources_max.nodes = 1:ppn=1
>>>> set queue serial resources_max.walltime = 1000:00:00
>>>> set queue serial resources_default.cput = 336:00:00
>>>> set queue serial resources_default.mem = 900mb
>>>> set queue serial resources_default.ncpus = 1
>>>> set queue serial resources_default.nodect = 1
>>>> set queue serial resources_default.nodes = 1:ppn=1
>>>> set queue serial enabled = True
>>>> set queue serial started = True
>>>> #
>>>> # Create and define queue workq
>>>> #
>>>> create queue workq
>>>> set queue workq queue_type = Execution
>>>> set queue workq resources_max.cput = 10000:00:00
>>>> set queue workq resources_max.ncpus = 200
>>>> set queue workq resources_max.nodect = 64
>>>> set queue workq resources_max.nodes = 200:ppn=4
>>>> set queue workq resources_max.walltime = 10000:00:00
>>>> set queue workq resources_min.cput = 00:00:01
>>>> set queue workq resources_min.ncpus = 1
>>>> set queue workq resources_min.nodect = 1
>>>> set queue workq resources_min.walltime = 00:00:01
>>>> set queue workq resources_default.cput = 10000:00:00
>>>> set queue workq resources_default.nodect = 1
>>>> set queue workq resources_default.walltime = 10000:00:00
>>>> set queue workq enabled = True
>>>> set queue workq started = True
>>>> #
>>>> # Create and define queue parallel
>>>> #
>>>> create queue parallel
>>>> set queue parallel queue_type = Execution
>>>> set queue parallel resources_max.cput = 10000:00:00
>>>> set queue parallel resources_max.ncpus = 200
>>>> set queue parallel resources_max.nodect = 64
>>>> set queue parallel resources_max.nodes = 200:ppn=4
>>>> set queue parallel resources_max.walltime = 10000:00:00
>>>> set queue parallel resources_min.ncpus = 1
>>>> set queue parallel resources_min.nodect = 1
>>>> set queue parallel resources_default.ncpus = 1
>>>> set queue parallel resources_default.nodect = 1
>>>> set queue parallel resources_default.nodes = 1:ppn=1
>>>> set queue parallel resources_default.walltime = 10000:00:00
>>>> set queue parallel enabled = True
>>>> set queue parallel started = True
>>>> #
>>>> # Set server attributes.
>>>> #
>>>> set server scheduling = True
>>>> set server acl_host_enable = False
>>>> set server acl_user_enable = False
>>>> set server default_queue = serial
>>>> set server log_events = 127
>>>> set server mail_from = adm
>>>> set server query_other_jobs = True
>>>> set server resources_available.ncpus = 200
>>>> set server resources_available.nodect = 64
>>>> set server resources_available.nodes = 200
>>>> set server resources_default.neednodes = 1
>>>> set server resources_default.nodect = 1
>>>> set server resources_default.nodes = 1
>>>> set server resources_max.ncpus = 200
>>>> set server resources_max.nodes = 200
>>>> set server scheduler_iteration = 60
>>>> set server node_check_rate = 150
>>>> set server tcp_timeout = 6
>>>> set server default_node = 1
>>>> set server pbs_version = 2.0.0p8
>>>>
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Nilesh Mistry
>>>> Academic Computing Services
>>>> [EMAIL PROTECTED] & TEL Campus
>>>> Seneca College Of Applies Arts & Technology
>>>> 70 The Pond Road
>>>> Toronto, Ontario
>>>> M3J 3M6 Canada
>>>> Phone 416 491 5050 ext 3788
>>>> Fax 416 661 4695
>>>> http://acs.senecac.on.ca
>>>>
>>>>
>>>>
>>>> Michael Edwards wrote:
>>>>
>>>>         
>>>>> We'd need your script and the qsub command you used, possibly more
>>>>> configuration information from maui and torque, to be much help.
>>>>>
>>>>> I don't know that we have anyone who is deep with maui or torque right
>>>>> now, you might also want to ask on the maui or torque lists.
>>>>>
>>>>> >From the other posts you have made this error seems to be one of those
>>>>> general "Something is Broken" messages that could have many causes.
>>>>>
>>>>> On 9/17/07, Nilesh Mistry <[EMAIL PROTECTED]> wrote:
>>>>>
>>>>>
>>>>>           
>>>>>> Hello
>>>>>>
>>>>>> I am having problems submitting job that requires 23 threads.  I keep
>>>>>> getting the following error:
>>>>>>
>>>>>> ERROR: Number of meshes not equal to number of thread
>>>>>>
>>>>>> Hardware:
>>>>>> 10 quad core nodes (therefore 40 processors available)
>>>>>>
>>>>>> What do I need to insure in my job queue (qmgr) , maui (maui.cfg) and
>>>>>> my submit script when using qsub?
>>>>>>
>>>>>> Any and all help is greatly appreciated.
>>>>>>
>>>>>> --
>>>>>> Thanks
>>>>>>
>>>>>> Nilesh Mistry
>>>>>> Academic Computing Services
>>>>>> [EMAIL PROTECTED] & TEL Campus
>>>>>> Seneca College Of Applies Arts & Technology
>>>>>> 70 The Pond Road
>>>>>> Toronto, Ontario
>>>>>> M3J 3M6 Canada
>>>>>> Phone 416 491 5050 ext 3788
>>>>>> Fax 416 661 4695
>>>>>> http://acs.senecac.on.ca
>>>>>>
>>>>>>
>>>>>>
>>>>>> -------------------------------------------------------------------------
>>>>>> This SF.net email is sponsored by: Microsoft
>>>>>> Defy all challenges. Microsoft(R) Visual Studio 2005.
>>>>>> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
>>>>>> _______________________________________________
>>>>>> Oscar-users mailing list
>>>>>> Oscar-users@lists.sourceforge.net
>>>>>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>>>>>
>>>>>>
>>>>>>
>>>>>>             
>>>>> -------------------------------------------------------------------------
>>>>> This SF.net email is sponsored by: Microsoft
>>>>> Defy all challenges. Microsoft(R) Visual Studio 2005.
>>>>> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
>>>>> _______________________________________________
>>>>> Oscar-users mailing list
>>>>> Oscar-users@lists.sourceforge.net
>>>>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>>>>
>>>>>
>>>>>           
>>>> -------------------------------------------------------------------------
>>>> This SF.net email is sponsored by: Microsoft
>>>> Defy all challenges. Microsoft(R) Visual Studio 2005.
>>>> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
>>>> _______________________________________________
>>>> Oscar-users mailing list
>>>> Oscar-users@lists.sourceforge.net
>>>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>>>
>>>>
>>>>         
>>> -------------------------------------------------------------------------
>>> This SF.net email is sponsored by: Microsoft
>>> Defy all challenges. Microsoft(R) Visual Studio 2005.
>>> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
>>> _______________________________________________
>>> Oscar-users mailing list
>>> Oscar-users@lists.sourceforge.net
>>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>>
>>>       
>> -------------------------------------------------------------------------
>> This SF.net email is sponsored by: Microsoft
>> Defy all challenges. Microsoft(R) Visual Studio 2005.
>> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
>> _______________________________________________
>> Oscar-users mailing list
>> Oscar-users@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>
>>     
>
> -------------------------------------------------------------------------
> This SF.net email is sponsored by: Microsoft
> Defy all challenges. Microsoft(R) Visual Studio 2005.
> http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
> _______________________________________________
> Oscar-users mailing list
> Oscar-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/oscar-users
>   


-------------------------------------------------------------------------
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2005.
http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
_______________________________________________
Oscar-users mailing list
Oscar-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/oscar-users

Reply via email to