Re: [gridengine users] Minimum number of slots

2018-02-01 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)

> #$ -l nodes=2,ppn=12  (--> here: 12 slots on 2 nodes = 24 in sum) .
>
> when you restrict "nodes=1" (with resource quotas "qconf -mrqs" for example) 
> then nobody > should be able to use more than 1 node.

sorry, correction: "to use more than 1 node PER JOB" of course.


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Minimum number of slots

2018-02-01 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi Ansger,

do you have experiences with Torque/PBS?

There is a useful gridengine plugin which works quite similar: 
https://github.com/brlindblom/gepetools

I have it in use on one of my clusters. The advantage: users must order not 
just slots but also nodes in the form:

#$ -l nodes=2,ppn=12  (--> here: 12 slots on 2 nodes = 24 in sum) .

when you restrict "nodes=1" (with resource quotas "qconf -mrqs" for example) 
then nobody should be able to use more than 1 node.

If you are interested and don't want to read all the gepetool instructions, I 
can send you my personal documentation.

Ursula


Von: users-boun...@gridengine.org  im Auftrag von 
Ansgar Esztermann-Kirchner 
Gesendet: Donnerstag, 01. Februar 2018 11:44
An: users@gridengine.org
Betreff: [gridengine users] Minimum number of slots

Hello List,

we're on 2011.11, and our general setup has nodes with a mixture of
CPUs (e.g. 8, 20, 40 cores). Most of the nodes lack a high-speed
interconnect, so we use a PE with allocation_rule $pe_slots, limiting
jobs to just a single machine. We're also using fairshare to achieve
an even distribution of resources to users in the long term.
There is a trade-off between fairshare and optimal resource usage when
only low-priority users have 40-core jobs and a 40-core node becomes
free. I know I can set my preferences by setting the relative weights
for fairshare and urgency.

Now, I think I can improve upon this choice by creating separate
queues for different machines "sizes", i.e. an 8-core queue, a
20-core queue and so on. However, I do not see a (tractable) way to
enforce proper job-queue association: allocation_rule 8 (etc) comes to
mind, but I would lose the crucial one-host limit. This could be
circumvented by creating one PE per node, but that would mean a huge
administrative burden (and possible also a lot of extra load on the
scheduler).

Anything I'm missing?
Thanks a lot,

A.
--
Ansgar Esztermann
Sysadmin
http://www.mpibpc.mpg.de/grubmueller/esztermann

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] CentOS 7 rpm package available?

2016-07-06 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hello Arnau,

As far as I know, there are no Centos 7 packages – but you can install the 
CentOS 6 rpm packages without problems.

Regards,
Ursula

Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im 
Auftrag von Arnau
Gesendet: Freitag, 01. Juli 2016 15:29
An: Gridengine Users Group
Betreff: [gridengine users] CentOS 7 rpm package available?

Hello all,

anyone knows if is there a rpm package for rh7 like distros  available?
anyone already buit it? could you please share it (or your build notes)?

TIA,
Arnau
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Suspension of jobs and consumeable resources

2015-11-16 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
>> Am 16.11.2015 um 11:41 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at) 
>> <ursula.wink...@uni-graz.at>:
>> 
>> Dear gridengine members,
>> 
>> when there are two queues and one is a subordinate of the other, then 
>> the release/shift of slots to the superordinate jobs is no problem, 
>> but (other) consumeable resources like h_vmem are. Does somebody have an 
>> idea to bypass that?
>
>As the memory is still occupied by the suspended job, I'm not sure whether 
>it's wise to bypass it. Even in case 
> the job starts, it may run out of memory at one time.
>
>Although I usually use only 2 GB of swap nowadays, you could try to have swap 
>as large as the built-in memory. 
> The result should be, that any suspended job can be swapped out completely.
>
> -- Reuti

So long I just lowered the default value of h_vmem (assigned when it isn't 
ordered explicit with qsub).
That works in some limited way because many users just overestimate the true 
memory demand of their jobs. 

But the idea with swapping out sounds at least worth to be considered in some 
cases. How do I implement it within the
gridengine?

Ursula

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


[gridengine users] Suspension of jobs and consumeable resources

2015-11-16 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Dear gridengine members,

when there are two queues and one is a subordinate of the other, then the 
release/shift
of slots to the superordinate jobs is no problem, but (other) consumeable 
resources
like h_vmem are. Does somebody have an idea to bypass that?

Thanks,
Ursula
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-31 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 So it sounds like the $pe_slots=serial behavior isn't true for your grid 
 engine.  
 An alternative would be to convert single node jobs to serial jobs that 
 requests 
 a consumable(other than slots)  representing the number of cores on the node. 
  
 However this will change the behavior when requesting other consumables and 
 probably won't interact well with core binding. 

I'm just testing another way/trick: I attached to each queue-instance for each 
host an own
seq_no (numbers increasing with +1). E.g.
# qconf -sq mpi
seq_no   0, [f01=101],[f02=102], ...

So far the tests look pretty good now! :-)



___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-31 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 What happens with serial jobs (ie no PE at all)?  If they exhibit the same 
 behavior then we 
 just need to figure out how to tweak the scheduler config to fill up nodes.  
 If serial jobs are 
 already clumping onto a few nodes then the claim that $pe_slots behaves like 
 serial on the 
 oracle blog that Reuti pointed to may not be an accurate description of your 
 version of grid engine.

Because gepetools is configured as default jsv ( jsv_url   
/software/sge/gepetools/pe.jsv ) for
all jobs, serial jobs are also concerned (nodes=1,ppn=1). 

Just for testing I disabled the jsv temporarily: the scheduling with serial 
jobs work then as in the 
oracle blog (= filling up) 

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-31 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)

 Have you got slots defined per host or just per queue?
 It may need to be defined at a host level via complex_values
 in order to work.

Both:

# qconf -sc | grep #name\|slots
#name   shortcut   typerelop   requestable consumable 
default  urgency 
occupied_slots  o_slotsINT =  YES YES1 
   0
slots   s  INT =  YES YES1 
   1000


# qconf -se f20
hostname  f20
load_scaling  NONE
complex_valuesoccupied_slots=12,slots=12,num_proc=12,h_vmem=45.01G, \
  exclusive=true

# qconf -sq mpi
   ...
   slots  1,[f01=12],[f02=12],[f38=12],[f03=12],[f04=12],[f05=12], \
  [f06=12],[f07=12],[f08=12],[f09=12],[f10=12],[f11=12], \
  [f12=12],[f13=12],[f14=12],[f15=12],[f16=12],[f17=12], \
  [f18=12],[f19=12],[f20=12],[f21=12],[f22=12],[f23=12], \
  [f25=12],[f26=12],[f27=12],[f28=12],[f29=12],[f30=12], \
  [f31=12],[f32=12],[f33=12],[f34=12],[f35=12],[f36=12], \
  [f37=12],[f24=12]
   ...

Maybe occupied_slots is the key???
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-31 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 complex_valuesoccupied_slots=12,slots=12,num_proc=12,h_vmem=45.01G, \
   exclusive=true

 To streamline this (but it won't solve your problem): I usually recommend not 
 to set num_proc, 
 as it's a feature of the host which is detected by SGE already. Limitations 
 are implemented by
 the slots complex already.

Then I'll remove it.

   slots  1,[f01=12],[f02=12],[f38=12],[f03=12],[f04=12],[f05=12], \
  [f06=12],[f07=12],[f08=12],[f09=12],[f10=12],[f11=12], \
  [f12=12],[f13=12],[f14=12],[f15=12],[f16=12],[f17=12], \
  [f18=12],[f19=12],[f20=12],[f21=12],[f22=12],[f23=12], \
  [f25=12],[f26=12],[f27=12],[f28=12],[f29=12],[f30=12], \
  [f31=12],[f32=12],[f33=12],[f34=12],[f35=12],[f36=12], \
  [f37=12],[f24=12]

 When all are having 12, it can just be put as a single value here.

Ok.

Thank you for your optimation-recommendations!

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-30 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 Sorry to step in the discussion: `qstat -j ...` shows the requested one, the 
 granted one is in `qstat -r`.

 $ qsub -pe * 2 test.sh
  Your job 44329 (test.sh) has been submitted $ qstat -j 44329 ...
 parallel environment:  * range: 2
 ...

My jobs:

qstat -j ...
   ...
   parallel environment: gepetools_1host range 2
   ...

That's the PE I created for that purposes.  So qstat -j shows the right info.


 $ qstat -r
 ...
Requested PE: * 2
Granted PE:   make 2

qstat -r
...
Requested PE:   gepetools_1host 2
Granted PE:gepetools_1host 2

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-30 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 Well, I created an additional PE with alloacation_rule $pe_slots, 
 and built in an if condition in pe.jsv for all jobs which request 
 just a single node to be assigned to this new PE. But the annoying 
 situation didn't change. The scheduler configuration is set to
 queue_sort_methodload and load_formula  slots. So what I'm
 still missing?
I believe it should be a load_formula of -slots so the more slots are 
available(fewest used) 
the lower the load and the more attractive the node.  
The page Reuti pointed to manages to write this both ways around.

Setting load_formula to -slots doesn't change anything - every job still 
starts on a separate host
(but in this case it should be the correct hehave if I don't misinterpret the 
instructions from the Web Page Reuti  mentioned).

I must be missing something else and pretty basic...

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-30 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)

 I believe it should be a load_formula of -slots so the more slots are
 available(fewest used) the lower the load and the more attractive the
 node.  The page Reuti pointed to manages to write this both ways around.

I'll try it out tomorrow - I'm not at the office now and it's a little bit 
difficult from here.
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-30 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)


 On Thu, 30 Jul 2015 12:57:13 +
 Winkler, Ursula (ursula.wink...@uni-graz.at)
 ursula.wink...@uni-graz.at wrote:
 
 My suggestion was to modify your jsv/gepetools to force single node
 parallel jobs into PEs with $pe_slots allocation rules (which gives
 you control over where they are scheduled via queue_sort_method and
 load_formula) while sending the others to PEs with other
 (appropriate) 
 allocation rules that won't cause (ii).
 
 Well, I created an additional PE with alloacation_rule $pe_slots,
 and built in an if condition in pe.jsv for all jobs which request
 just a single node to be assigned to this new PE. But the annoying
 situation didn't change. The scheduler configuration is set to
 queue_sort_methodload and load_formula  slots. So what I'm
 still missing?
 Ignore previous message.  Me getting it back to front I think.  That
 looks correct (I think).  Have you checked the jobs show the right
 granted PE with qstat -j?

Yes, of course.

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Filling up nodes when using gepetools

2015-07-30 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)




Am 30.07.2015 um 18:29 schrieb Reuti re...@staff.uni-marburg.de:

 
 Am 30.07.2015 um 18:14 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at) 
 ursula.wink...@uni-graz.at:
 
 
 
 On Thu, 30 Jul 2015 12:57:13 +
 Winkler, Ursula (ursula.wink...@uni-graz.at)
 ursula.wink...@uni-graz.at wrote:
 
 My suggestion was to modify your jsv/gepetools to force single node
 parallel jobs into PEs with $pe_slots allocation rules (which gives
 you control over where they are scheduled via queue_sort_method and
 load_formula) while sending the others to PEs with other
 (appropriate) 
 allocation rules that won't cause (ii).
 
 Well, I created an additional PE with alloacation_rule $pe_slots,
 and built in an if condition in pe.jsv for all jobs which request
 just a single node to be assigned to this new PE. But the annoying
 situation didn't change. The scheduler configuration is set to
 queue_sort_methodload and load_formula  slots. So what I'm
 still missing?
 Ignore previous message.  Me getting it back to front I think.  That
 looks correct (I think).  Have you checked the jobs show the right
 granted PE with qstat -j?
 
 Yes, of course.
 
 Sorry to step in the discussion: `qstat -j ...` shows the requested one, the 
 granted one is in `qstat -r`.
 
 $ qsub -pe * 2 test.sh
 Your job 44329 (test.sh) has been submitted
 $ qstat -j 44329
 ...
 parallel environment:  * range: 2
 ...
 $ qstat -r
 ...
   Requested PE: * 2
   Granted PE:   make 2
 
 -- Reuti

At the moment I don't know if I checked it with qstat -j, but I checked it - 
when I'm in the office again I probably have the output still on some screen 
window so I can tell it exactly. And I did do a test: I removed the PE 
temporarely from the queue - with the result that the jobs could not start 
anymore (as respected). 
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-23 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi Reuti,

thank you for asking. The answer is unfortunately not. But at least the 
calculation is correct (qstat -g c), as soon as the user requests not all of 
the available slots.
So the unsatisfying solution is that he's submitting his batch jobs now with 
less slots...

Ursula

-Ursprüngliche Nachricht-
Von: Reuti [mailto:re...@staff.uni-marburg.de] 
Gesendet: Donnerstag, 23. Oktober 2014 11:26
An: Winkler, Ursula (ursula.wink...@uni-graz.at)
Cc: William Hay; Gridengine Users Group
Betreff: Re: AW: [gridengine users] Weird scheduler calculation error?

Hi,

Am 21.10.2014 um 14:23 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at):

 schedd_job_info is switched on then? But even if switched off it should 
 show up in `qalter -w p`.
 
 Yes, it is on.
 
 And 24 slots per machine then - `qstat -g c ` reveals the slots as being 
 free too?
 
 A good question: it reveals that from 72 are now (!) 70 are free and no one 
 is used. What's happening here?

Did you get any further insight about this issue?

-- Reuti


 Ursula
 
 And there is just that job and no other ones.
 
 
 Ursula
 
 ___
 users mailing list
 users@gridengine.org
 https://gridengine.org/mailman/listinfo/users
 
 


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-23 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)

 What is the setting of slots in the PE itself?

492(the cluster has 35 hosts with 12 slots and 3 hosts with 24 slots; but 
all hosts with 12 slots are disabled in this  testqueue).

Ursula


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-21 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi Reuti,

no - and there is no (other than reputed slots) resource shortage. And no host 
is in error state.

Ursula


-Ursprüngliche Nachricht-
Von: Reuti [mailto:re...@staff.uni-marburg.de] 
Gesendet: Dienstag, 21. Oktober 2014 12:25
An: Gridengine Users Group
Cc: Winkler, Ursula (ursula.wink...@uni-graz.at)
Betreff: Re: [gridengine users] Weird scheduler calculation error?

Hi,

Am 21.10.2014 um 11:21 schrieb Ursula Winkler:

 Hi gridengine members,
 
 For now I ran out of ideas with an annoying problem:
 
 A job with 72 slots does not start because of qstat -j jobid tells 
 cannot run in PE mpios because it only offers 64 slots, but there are 72 
 free (qalter -w p jobid and qalter -w -v jobid tells  verification: 
 found possible assignment with 72 slots and nothing more). The PE and host 
 configurations are ok. I already restarted the master-process and the 
 execution-daemons on the concerning hosts, but (as expected) that didn't help.

Any calendar in effect?

-- Reuti


 So what could be wrong? Any suggestions are appreciated.
 
 Ursula
 
 ___
 users mailing list
 users@gridengine.org
 https://gridengine.org/mailman/listinfo/users


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-21 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
The allocation rule is $fill_up and not $pe_slots. So that should be ok.

Ursula

-Ursprüngliche Nachricht-
Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im 
Auftrag von Simon Andrews
Gesendet: Dienstag, 21. Oktober 2014 13:05
An: Gridengine Users Group
Betreff: Re: [gridengine users] Weird scheduler calculation error?

What is your allocation rule for the mpios parallel environment (qconf -sp 
mpios)?  Could it be that the allocation says that the slots all have to be on 
the same physical node, and no single node has more than 64 slots available?


-Original Message-
From: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] On 
Behalf Of Winkler, Ursula (ursula.wink...@uni-graz.at)
Sent: 21 October 2014 11:46
To: Gridengine Users Group
Subject: Re: [gridengine users] Weird scheduler calculation error?

Hi Reuti,

no - and there is no (other than reputed slots) resource shortage. And no host 
is in error state.

Ursula


-Ursprüngliche Nachricht-
Von: Reuti [mailto:re...@staff.uni-marburg.de]
Gesendet: Dienstag, 21. Oktober 2014 12:25
An: Gridengine Users Group
Cc: Winkler, Ursula (ursula.wink...@uni-graz.at)
Betreff: Re: [gridengine users] Weird scheduler calculation error?

Hi,

Am 21.10.2014 um 11:21 schrieb Ursula Winkler:

 Hi gridengine members,

 For now I ran out of ideas with an annoying problem:

 A job with 72 slots does not start because of qstat -j jobid tells 
 cannot run in PE mpios because it only offers 64 slots, but there are 72 
 free (qalter -w p jobid and qalter -w -v jobid tells  verification: 
 found possible assignment with 72 slots and nothing more). The PE and host 
 configurations are ok. I already restarted the master-process and the 
 execution-daemons on the concerning hosts, but (as expected) that didn't help.

Any calendar in effect?

-- Reuti


 So what could be wrong? Any suggestions are appreciated.

 Ursula

 ___
 users mailing list
 users@gridengine.org
 https://gridengine.org/mailman/listinfo/users


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
The Babraham Institute, Babraham Research Campus, Cambridge CB22 3AT Registered 
Charity No. 1053902.
The information transmitted in this email is directed only to the addressee. If 
you received this in error, please contact the sender and delete this email 
from your system. The contents of this e-mail are the views of the sender and 
do not necessarily represent the views of the Babraham Institute. Full 
conditions at: www.babraham.ac.ukhttp://www.babraham.ac.uk/terms

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-21 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)

 The qsub man page states that -w p and -w v don't take into account load 
 values.  Possibly the job is requesting a complex whose value is determined 
 by a load sensor and the returned value is not suitable but not causing an 
 alarm.

Should not qstat -j jobid list the shortage of a complex? There's nothing 
(I already checked the job script of the user too). And the hosts on which the 
job should run are completely free.
 
 Alternatively perhaps a higher priority job with a resource reservation in 
 place?  We run with schedd_job_info false so I don't have much experience 
 with it but I would expect qstat -j to take reservations into account while 
 qalter -w [pv] can't as they're basically just doing the calculation for that 
 one job.

No.  It's a testqueue with 3 hosts reserved only for that queue. And there is 
just that job and no other ones.


Ursula

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Weird scheduler calculation error?

2014-10-21 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 schedd_job_info is switched on then? But even if switched off it should 
 show up in `qalter -w p`.

Yes, it is on.

 And 24 slots per machine then - `qstat -g c ` reveals the slots as being free 
 too?

A good question: it reveals that from 72 are now (!) 70 are free and no one is 
used. What's happening here?

Ursula

 And there is just that job and no other ones.
 
 
 Ursula
 
 ___
 users mailing list
 users@gridengine.org
 https://gridengine.org/mailman/listinfo/users


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] Hosts not fully used with fill up

2014-09-18 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi Rafa,

for such purposes we have configured a separate Parallel Environment with 
allocation_rule $pe_slots (instead of $fill_up). Jobs scheduled with this 
rule can run ONLY on one host.

Regards,
Usula

-Ursprüngliche Nachricht-
Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im 
Auftrag von Rafael Arco Arredondo
Gesendet: Donnerstag, 18. September 2014 09:44
An: users@gridengine.org
Betreff: [gridengine users] Hosts not fully used with fill up

Hello everyone,

We are having an issue with the parallel environments and the allocation of 
slots with the fill up policy.

Although we have configured the resource quotas of the queues not to use more 
than the number of slots the machine have and we control in the prolog that the 
jobs be submitted with a number of slots multiple of the number of physical 
processors, we are observing that sometimes, the slots of a job are split into 
several nodes, when they should be running in only one node.

We are using Open Grid Scheduler 2011.11p1. This didn't happen in SGE 6.2.

Has anyone experienced the same situation? Any clues of why it is happening?

Thanks in advance,

Rafa

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] submitting to a specific host

2014-03-21 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi, try

qsub -q queue@host,... (e.g. qsub -q mpi@host1,mpi@host2, )

Ursula

Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im 
Auftrag von Henrichs, Juryk
Gesendet: Freitag, 21. März 2014 11:24
An: users@gridengine.org
Betreff: [gridengine users] submitting to a specific host

Hi,

We are using SunGridEngine 6.5-2.

To submit a job to a specific host on our cluster. I tried something like:
qsub -l hostname=nodeXXX jobfile.sge

Unfortunately the job does not get started although the nodeXXX is active in 
the requested queue and empty.

qstat -j seems not to give a useful hint. However, the last two lines are:
cannot run because it exceeds limit / in rule isis_licenses/1
cannot run in PE peSlots because it only offers 0 slots

The resource isis_licenses is not requested here, so that shouldn't be a 
problem here?

Any suggestions at where to start searching the problem?

Juryk
This e-mail and any attachment thereto may contain confidential information 
and/or information protected by intellectual property rights for the exclusive 
attention of the intended addressees named above. Any access of third parties 
to this e-mail is unauthorised. Any use of this e-mail by unintended recipients 
such as total or partial copying, distribution, disclosure etc. is prohibited 
and may be unlawful. When addressed to our clients the content of this e-mail 
is subject to the General Terms and Conditions of DNV GL Group of Companies 
applicable at the date of this e-mail.
If you have received this e-mail in error, please notify the sender either by 
telephone or by e-mail and delete the material from any computer.
DNV GL Group of Companies does not warrant and/or guarantee that this message 
at the moment of receipt is authentic, correct and its communication free of 
errors, interruption etc.
FutureShip GmbH, HRB 106781 AG HH, VAT Reg. No. DE263937825
Geschäftsführer (CEO): Volker Höppner, Stefan Deucker
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


[gridengine users] qacct and resource requests

2014-01-20 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Dear gridengine list members,

I've a question concerning qacct and resource requests. qacct prints out a 
lot of information - but little about resource requests that jobs asked for by 
qsub -l 
Of course it is possible to find out via qacct -l the jobs with a 
*particular* value (as one has to specify e.g. qacct -l  h_vmem=2.5G ...) but 
there doesn't seem to exist a possibility to find out in a more direct way the 
values jobs have ordered (I need it for statistical purposes).

Would be great if somebody has an idea.

Cheers, Ursula

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] qacct and resource requests

2014-01-20 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 I've a question concerning qacct and resource requests. qacct prints out 
 a lot of information - but little about resource requests that jobs asked 
 for by qsub -l 
 Of course it is possible to find out via qacct -l the jobs with a 
 *particular* value (as one has to specify e.g. qacct -l  h_vmem=2.5G ...) 
 but there doesn't seem to exist a possibility to find out in a more direct 
 way the values jobs have ordered (I need it for statistical purposes).

They are recorded in the accounting file in plain text in the field 
category (`man accounting`).

-- Reuti

Thank you, Reuti!

Ursula

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


[gridengine users] Jobs being killed when another jobs start

2012-05-04 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Hi everyone,

I have a problem with SGE: it looks like (though I'm really not sure) that 
running jobs are killed when other jobs are submitted which otherwise should 
wait in the queue because no free cores are available. It happens repeatedly so 
this raises suspicion. Unfortunately the error of the job log files only says  
that the jobs terminated badly. qacct -j job-no tells that the exit status 
of the jobs is 7. I could not find out what this error code means. Does 
anybody know that?

Thanks, Ursula

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] qtcsh building error

2012-03-29 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Oh no, sorry, I was too hasty: adding  -lcurses to

  gcc -o tcsh - ... -DPROG_NAME=qtcsh  -lcrypt

of course produced just tcsh, not qtcsh.  Now, can I rename it to qtcsh, or 
what's about qtcsh?

Ursula

Von: Tim Wright [mailto:tim.dormo...@gmail.com]
Gesendet: Mittwoch, 28. März 2012 17:22
An: Winkler, Ursula (ursula.wink...@uni-graz.at)
Betreff: Re: [gridengine users] qtcsh building error

Have you tried adding -lcurses or -lncurses to your compile string?
On 28 March 2012 15:23, Ursula Winkler 
ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at wrote:
undefined reference to `tputs'

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] qtcsh building error

2012-03-29 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
Thank you!

Ursula

Von: Ron Chen [mailto:ron_chen_...@yahoo.com]
Gesendet: Donnerstag, 29. März 2012 09:34
An: Winkler, Ursula (ursula.wink...@uni-graz.at); 'Tim Wright'
Cc: 'users@gridengine.org'
Betreff: Re: [gridengine users] qtcsh building error

The install script copies tcsh to qtcsh into the final installable package. So 
you don't need to worry about not getting qtcsh.

 -Ron


From: Winkler, Ursula 
(ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at) 
ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at
To: 'Tim Wright' tim.dormo...@gmail.commailto:tim.dormo...@gmail.com
Cc: 'users@gridengine.org' users@gridengine.orgmailto:users@gridengine.org
Sent: Thursday, March 29, 2012 3:28 AM
Subject: Re: [gridengine users] qtcsh building error

Oh no, sorry, I was too hasty: adding  “-lcurses” to

  “gcc –o tcsh - … -DPROG_NAME=”qtcsh” …. –lcrypt”

of course produced just “tcsh”, not “qtcsh”.  Now, can I rename it to qtcsh, or 
what’s about qtcsh?

Ursula

Von: Tim Wright [mailto:tim.dormo...@gmail.com]
Gesendet: Mittwoch, 28. März 2012 17:22
An: Winkler, Ursula 
(ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at)
Betreff: Re: [gridengine users] qtcsh building error

Have you tried adding -lcurses or -lncurses to your compile string?
On 28 March 2012 15:23, Ursula Winkler 
ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at wrote:
undefined reference to `tputs'


___
users mailing list
users@gridengine.orgmailto:users@gridengine.org
https://gridengine.org/mailman/listinfo/users

___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users


Re: [gridengine users] qtcsh building error

2012-03-28 Thread Winkler, Ursula (ursula.wink...@uni-graz.at)
 I know it's possible to build the binaries without qtcsh but I'm not sure if 
 it is not needed later? Can I disclaim it?

qtcsh is a version of tcsh where certain commands (which you can specify) are 
send to a node by SGE instead of being executed on the machine where you 
issued the command. If you don't use tcsh, you most likely won't need it.

 http://arc.liv.ac.uk/SGE/htmlman/htmlman1/qtcsh.html

-- Reuti

Then it's not so fine disclaiming it because I have users who use tcsh. 
Is there a workaround to get the compilation error solved?

Ursula


___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users