On Mon, Jan 22, 2007 at 08:10:01PM -0600, Michael Edwards wrote:
>I suspect that the scripts which OSCAR runs on detect the cpu count
>but do not detect multiple cores as separate "processors".  If this is
>the case, this is a fairly major issue and needs to be addressed soon.
> Multi-core processors are becoming more and more common.  I don't
>have any hardware to test this on myself yet, but I should in a couple
>months.

oscar5 x86_64 centos detected our dual dual-core's just fine.
I think it set this up in the 'complete install' thingy.

also you can always just edit the right amount of cores into
/var/spool/pbs/server_priv/nodes and then restart pbs_server

cheers,
robin

>
>On the plus side, it ought to be a fairly easy fix...
>
>On 22 Jan 2007 20:37:59 -0500, [EMAIL PROTECTED]
><[EMAIL PROTECTED]> wrote:
>> Thanks for the message!
>>
>> Actually, I only have two client nodes. One have 4 dual-core AMD CPUs
>> (total 8 logic CPUs). The other host has 2 dual-core CPUs (total 4 logic
>> CPUs).
>>
>> I use "qsub ./script.sh" to submit jobs. All the jobs use the same script.
>>
>> If I submit 12 jobs, only 6 jobs runs (4 on the first node and 2 on the
>> other node) at the same time.
>>
>> Here is what I get if I do "print server" from qmgr. Do I need to change
>> anything?
>>
>> ======================================================
>>
>> #
>> # Create queues and set their attributes.
>> #
>> #
>> # Create and define queue workq
>> #
>> create queue workq
>> set queue workq queue_type = Execution
>> set queue workq resources_max.cput = 10000:00:00
>> set queue workq resources_max.ncpus = 12
>> set queue workq resources_max.nodect = 2
>> set queue workq resources_max.walltime = 10000:00:00
>> set queue workq resources_min.cput = 00:00:01
>> set queue workq resources_min.ncpus = 1
>> set queue workq resources_min.nodect = 1
>> set queue workq resources_min.walltime = 00:00:01
>> set queue workq resources_default.cput = 10000:00:00
>> set queue workq resources_default.ncpus = 1
>> set queue workq resources_default.nodect = 1
>> set queue workq resources_default.walltime = 10000:00:00
>> set queue workq resources_available.nodect = 2
>> set queue workq enabled = True
>> set queue workq started = True
>> #
>> # Set server attributes.
>> #
>> set server scheduling = True
>> set server default_queue = workq
>> set server log_events = 64
>> set server mail_from = adm
>> set server query_other_jobs = True
>> set server resources_available.ncpus = 12
>> set server resources_available.nodect = 2
>> set server resources_available.nodes = 2
>> set server resources_max.ncpus = 12
>> set server resources_max.nodes = 2
>> set server scheduler_iteration = 60
>> set server node_check_rate = 150
>> set server tcp_timeout = 6
>> set server pbs_version = 2.0.0p8
>>
>> ===========================================================
>>
>>
>> Here is maui.cfg file
>>
>> ===========================================================
>>
>> # maui.cfg 3.2.6p14
>>
>> SERVERHOST              photon.bwh.harvard.edu
>> # primary admin must be first in list
>> ADMIN1                root
>>
>> # Resource Manager Definition
>>
>> RMCFG[DUAL.EFOCHT.DE] TYPE=PBS
>>
>> # Allocation Manager Definition
>>
>> AMCFG[bank]  TYPE=NONE
>>
>> # full parameter docs at
>> http://clusterresources.com/mauidocs/a.fparameters.html
>> # use the 'schedctl -l' command to display current configuration
>>
>> RMPOLLINTERVAL  00:00:10
>>
>> SERVERPORT            42559
>> SERVERMODE            NORMAL
>>
>> # Admin: http://clusterresources.com/mauidocs/a.esecurity.html
>>
>>
>> LOGFILE               maui.log
>> LOGFILEMAXSIZE        10000000
>> LOGLEVEL              3
>>
>> # Job Priority:
>> http://clusterresources.com/mauidocs/5.1jobprioritization.html
>>
>> QUEUETIMEWEIGHT       1
>>
>> # FairShare: http://clusterresources.com/mauidocs/6.3fairshare.html
>>
>> #FSPOLICY              PSDEDICATED
>> #FSDEPTH               7
>> #FSINTERVAL            86400
>> #FSDECAY               0.80
>>
>> # Throttling Policies:
>> http://clusterresources.com/mauidocs/6.2throttlingpolicies.html
>>
>> # NONE SPECIFIED
>>
>> # Backfill: http://clusterresources.com/mauidocs/8.2backfill.html
>>
>> BACKFILLPOLICY  ON
>> RESERVATIONPOLICY     CURRENTHIGHEST
>>
>> # Node Allocation:
>> http://clusterresources.com/mauidocs/5.2nodeallocation.html
>>
>> NODEALLOCATIONPOLICY  MINRESOURCE
>>
>> # QOS: http://clusterresources.com/mauidocs/7.3qos.html
>>
>> # QOSCFG[hi]  PRIORITY=100 XFTARGET=100 FLAGS=PREEMPTOR:IGNMAXJOB
>> # QOSCFG[low] PRIORITY=-1000 FLAGS=PREEMPTEE
>>
>> # Standing Reservations:
>> http://clusterresources.com/mauidocs/7.1.3standingreservations.html
>>
>> # SRSTARTTIME[test] 8:00:00
>> # SRENDTIME[test]   17:00:00
>> # SRDAYS[test]      MON TUE WED THU FRI
>> # SRTASKCOUNT[test] 20
>> # SRMAXTIME[test]   0:30:00
>>
>> # Creds: http://clusterresources.com/mauidocs/6.1fairnessoverview.html
>>
>> # USERCFG[DEFAULT]      FSTARGET=25.0
>> # USERCFG[john]         PRIORITY=100  FSTARGET=10.0-
>> # GROUPCFG[staff]       PRIORITY=1000 QLIST=hi:low QDEF=hi
>> # CLASSCFG[batch]       FLAGS=PREEMPTEE
>> # CLASSCFG[interactive] FLAGS=PREEMPTOR
>>
>> NODEACCESSPOLICY
>>
>> =====================================================================
>>
>>
>>
>> > Check ganglia at http://localhost/ganglia and see where those 6 jobs
>> > are.  Make sure in particular they are not all sitting on one node or
>> > something silly.  If you have 6 nodes and they are one per node, then
>> > the queue is probably set up to reserve an entire node for each
>> > process.  There is a flag in the torque config file (I think) that
>> > tells it to do this.
>> >
>> > Could you post the script you are queueing with and the qsub command
>> > you use to submit the job?
>> >
>> > Are you running an smp kernel on your head node, I assume?  If you
>> > happened to be running in the non-smp version when you installed
>> > torque/maui probably don't know that there is more than one processor
>> > available...
>> >
>> > Hopefully this gives you some thoughts as to where to start looking...
>> >
>> > On 22 Jan 2007 13:13:02 -0500, Jinsong Ouyang <[EMAIL PROTECTED]>
>> > wrote:
>> >>
>> >>
>> >>
>> >>
>> >> I am using OSCAR 5.0 & Fedora 5.0 x86_64.  I have total 12 logic CPUs on
>> >> computing nodes. I use qsub to submit jobs and can only have maximum 6
>> >> jobs
>> >> running simultaneously. Half of the CPUs are not used. Could anyone
>> >> please
>> >> tell me how to increase the number of running jobs? I tried to set
>> >> max_running using qmgr. It does not seem to change anything. Do I need
>> >> to
>> >> change anything in maui.cfg?
>> >>
>> >>
>> >>
>> >> Many thanks,
>> >>
>> >>
>> >>
>> >> JO
>> >> -------------------------------------------------------------------------
>> >> Take Surveys. Earn Cash. Influence the Future of IT
>> >> Join SourceForge.net's Techsay panel and you'll get the chance to share
>> >> your
>> >> opinions on IT & business topics through brief surveys - and earn cash
>> >> http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
>> >>
>> >> _______________________________________________
>> >> Oscar-users mailing list
>> >> Oscar-users@lists.sourceforge.net
>> >> https://lists.sourceforge.net/lists/listinfo/oscar-users
>> >>
>> >>
>> >>
>> >
>> > -------------------------------------------------------------------------
>> > Take Surveys. Earn Cash. Influence the Future of IT
>> > Join SourceForge.net's Techsay panel and you'll get the chance to share
>> > your
>> > opinions on IT & business topics through brief surveys - and earn cash
>> > http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
>> > _______________________________________________
>> > Oscar-users mailing list
>> > Oscar-users@lists.sourceforge.net
>> > https://lists.sourceforge.net/lists/listinfo/oscar-users
>> >
>>
>>
>>
>> -------------------------------------------------------------------------
>> Take Surveys. Earn Cash. Influence the Future of IT
>> Join SourceForge.net's Techsay panel and you'll get the chance to share your
>> opinions on IT & business topics through brief surveys - and earn cash
>> http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
>> _______________________________________________
>> Oscar-users mailing list
>> Oscar-users@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/oscar-users
>>
>
>-------------------------------------------------------------------------
>Take Surveys. Earn Cash. Influence the Future of IT
>Join SourceForge.net's Techsay panel and you'll get the chance to share your
>opinions on IT & business topics through brief surveys - and earn cash
>http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
>_______________________________________________
>Oscar-users mailing list
>Oscar-users@lists.sourceforge.net
>https://lists.sourceforge.net/lists/listinfo/oscar-users

-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
Oscar-users mailing list
Oscar-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/oscar-users

Reply via email to