Re: [gridengine users] Minimum number of slots
> #$ -l nodes=2,ppn=12 (--> here: 12 slots on 2 nodes = 24 in sum) . > > when you restrict "nodes=1" (with resource quotas "qconf -mrqs" for example) > then nobody > should be able to use more than 1 node. sorry, correction: "to use more than 1 node PER JOB" of course. ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Minimum number of slots
Hi Ansger, do you have experiences with Torque/PBS? There is a useful gridengine plugin which works quite similar: https://github.com/brlindblom/gepetools I have it in use on one of my clusters. The advantage: users must order not just slots but also nodes in the form: #$ -l nodes=2,ppn=12 (--> here: 12 slots on 2 nodes = 24 in sum) . when you restrict "nodes=1" (with resource quotas "qconf -mrqs" for example) then nobody should be able to use more than 1 node. If you are interested and don't want to read all the gepetool instructions, I can send you my personal documentation. Ursula Von: users-boun...@gridengine.orgim Auftrag von Ansgar Esztermann-Kirchner Gesendet: Donnerstag, 01. Februar 2018 11:44 An: users@gridengine.org Betreff: [gridengine users] Minimum number of slots Hello List, we're on 2011.11, and our general setup has nodes with a mixture of CPUs (e.g. 8, 20, 40 cores). Most of the nodes lack a high-speed interconnect, so we use a PE with allocation_rule $pe_slots, limiting jobs to just a single machine. We're also using fairshare to achieve an even distribution of resources to users in the long term. There is a trade-off between fairshare and optimal resource usage when only low-priority users have 40-core jobs and a 40-core node becomes free. I know I can set my preferences by setting the relative weights for fairshare and urgency. Now, I think I can improve upon this choice by creating separate queues for different machines "sizes", i.e. an 8-core queue, a 20-core queue and so on. However, I do not see a (tractable) way to enforce proper job-queue association: allocation_rule 8 (etc) comes to mind, but I would lose the crucial one-host limit. This could be circumvented by creating one PE per node, but that would mean a huge administrative burden (and possible also a lot of extra load on the scheduler). Anything I'm missing? Thanks a lot, A. -- Ansgar Esztermann Sysadmin http://www.mpibpc.mpg.de/grubmueller/esztermann ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] CentOS 7 rpm package available?
Hello Arnau, As far as I know, there are no Centos 7 packages – but you can install the CentOS 6 rpm packages without problems. Regards, Ursula Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im Auftrag von Arnau Gesendet: Freitag, 01. Juli 2016 15:29 An: Gridengine Users Group Betreff: [gridengine users] CentOS 7 rpm package available? Hello all, anyone knows if is there a rpm package for rh7 like distros available? anyone already buit it? could you please share it (or your build notes)? TIA, Arnau ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Suspension of jobs and consumeable resources
>> Am 16.11.2015 um 11:41 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at) >> <ursula.wink...@uni-graz.at>: >> >> Dear gridengine members, >> >> when there are two queues and one is a subordinate of the other, then >> the release/shift of slots to the superordinate jobs is no problem, >> but (other) consumeable resources like h_vmem are. Does somebody have an >> idea to bypass that? > >As the memory is still occupied by the suspended job, I'm not sure whether >it's wise to bypass it. Even in case > the job starts, it may run out of memory at one time. > >Although I usually use only 2 GB of swap nowadays, you could try to have swap >as large as the built-in memory. > The result should be, that any suspended job can be swapped out completely. > > -- Reuti So long I just lowered the default value of h_vmem (assigned when it isn't ordered explicit with qsub). That works in some limited way because many users just overestimate the true memory demand of their jobs. But the idea with swapping out sounds at least worth to be considered in some cases. How do I implement it within the gridengine? Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
[gridengine users] Suspension of jobs and consumeable resources
Dear gridengine members, when there are two queues and one is a subordinate of the other, then the release/shift of slots to the superordinate jobs is no problem, but (other) consumeable resources like h_vmem are. Does somebody have an idea to bypass that? Thanks, Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
So it sounds like the $pe_slots=serial behavior isn't true for your grid engine. An alternative would be to convert single node jobs to serial jobs that requests a consumable(other than slots) representing the number of cores on the node. However this will change the behavior when requesting other consumables and probably won't interact well with core binding. I'm just testing another way/trick: I attached to each queue-instance for each host an own seq_no (numbers increasing with +1). E.g. # qconf -sq mpi seq_no 0, [f01=101],[f02=102], ... So far the tests look pretty good now! :-) ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
What happens with serial jobs (ie no PE at all)? If they exhibit the same behavior then we just need to figure out how to tweak the scheduler config to fill up nodes. If serial jobs are already clumping onto a few nodes then the claim that $pe_slots behaves like serial on the oracle blog that Reuti pointed to may not be an accurate description of your version of grid engine. Because gepetools is configured as default jsv ( jsv_url /software/sge/gepetools/pe.jsv ) for all jobs, serial jobs are also concerned (nodes=1,ppn=1). Just for testing I disabled the jsv temporarily: the scheduling with serial jobs work then as in the oracle blog (= filling up) ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
Have you got slots defined per host or just per queue? It may need to be defined at a host level via complex_values in order to work. Both: # qconf -sc | grep #name\|slots #name shortcut typerelop requestable consumable default urgency occupied_slots o_slotsINT = YES YES1 0 slots s INT = YES YES1 1000 # qconf -se f20 hostname f20 load_scaling NONE complex_valuesoccupied_slots=12,slots=12,num_proc=12,h_vmem=45.01G, \ exclusive=true # qconf -sq mpi ... slots 1,[f01=12],[f02=12],[f38=12],[f03=12],[f04=12],[f05=12], \ [f06=12],[f07=12],[f08=12],[f09=12],[f10=12],[f11=12], \ [f12=12],[f13=12],[f14=12],[f15=12],[f16=12],[f17=12], \ [f18=12],[f19=12],[f20=12],[f21=12],[f22=12],[f23=12], \ [f25=12],[f26=12],[f27=12],[f28=12],[f29=12],[f30=12], \ [f31=12],[f32=12],[f33=12],[f34=12],[f35=12],[f36=12], \ [f37=12],[f24=12] ... Maybe occupied_slots is the key??? ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
complex_valuesoccupied_slots=12,slots=12,num_proc=12,h_vmem=45.01G, \ exclusive=true To streamline this (but it won't solve your problem): I usually recommend not to set num_proc, as it's a feature of the host which is detected by SGE already. Limitations are implemented by the slots complex already. Then I'll remove it. slots 1,[f01=12],[f02=12],[f38=12],[f03=12],[f04=12],[f05=12], \ [f06=12],[f07=12],[f08=12],[f09=12],[f10=12],[f11=12], \ [f12=12],[f13=12],[f14=12],[f15=12],[f16=12],[f17=12], \ [f18=12],[f19=12],[f20=12],[f21=12],[f22=12],[f23=12], \ [f25=12],[f26=12],[f27=12],[f28=12],[f29=12],[f30=12], \ [f31=12],[f32=12],[f33=12],[f34=12],[f35=12],[f36=12], \ [f37=12],[f24=12] When all are having 12, it can just be put as a single value here. Ok. Thank you for your optimation-recommendations! ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
Sorry to step in the discussion: `qstat -j ...` shows the requested one, the granted one is in `qstat -r`. $ qsub -pe * 2 test.sh Your job 44329 (test.sh) has been submitted $ qstat -j 44329 ... parallel environment: * range: 2 ... My jobs: qstat -j ... ... parallel environment: gepetools_1host range 2 ... That's the PE I created for that purposes. So qstat -j shows the right info. $ qstat -r ... Requested PE: * 2 Granted PE: make 2 qstat -r ... Requested PE: gepetools_1host 2 Granted PE:gepetools_1host 2 ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
Well, I created an additional PE with alloacation_rule $pe_slots, and built in an if condition in pe.jsv for all jobs which request just a single node to be assigned to this new PE. But the annoying situation didn't change. The scheduler configuration is set to queue_sort_methodload and load_formula slots. So what I'm still missing? I believe it should be a load_formula of -slots so the more slots are available(fewest used) the lower the load and the more attractive the node. The page Reuti pointed to manages to write this both ways around. Setting load_formula to -slots doesn't change anything - every job still starts on a separate host (but in this case it should be the correct hehave if I don't misinterpret the instructions from the Web Page Reuti mentioned). I must be missing something else and pretty basic... ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
I believe it should be a load_formula of -slots so the more slots are available(fewest used) the lower the load and the more attractive the node. The page Reuti pointed to manages to write this both ways around. I'll try it out tomorrow - I'm not at the office now and it's a little bit difficult from here. ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
On Thu, 30 Jul 2015 12:57:13 + Winkler, Ursula (ursula.wink...@uni-graz.at) ursula.wink...@uni-graz.at wrote: My suggestion was to modify your jsv/gepetools to force single node parallel jobs into PEs with $pe_slots allocation rules (which gives you control over where they are scheduled via queue_sort_method and load_formula) while sending the others to PEs with other (appropriate) allocation rules that won't cause (ii). Well, I created an additional PE with alloacation_rule $pe_slots, and built in an if condition in pe.jsv for all jobs which request just a single node to be assigned to this new PE. But the annoying situation didn't change. The scheduler configuration is set to queue_sort_methodload and load_formula slots. So what I'm still missing? Ignore previous message. Me getting it back to front I think. That looks correct (I think). Have you checked the jobs show the right granted PE with qstat -j? Yes, of course. ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Filling up nodes when using gepetools
Am 30.07.2015 um 18:29 schrieb Reuti re...@staff.uni-marburg.de: Am 30.07.2015 um 18:14 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at) ursula.wink...@uni-graz.at: On Thu, 30 Jul 2015 12:57:13 + Winkler, Ursula (ursula.wink...@uni-graz.at) ursula.wink...@uni-graz.at wrote: My suggestion was to modify your jsv/gepetools to force single node parallel jobs into PEs with $pe_slots allocation rules (which gives you control over where they are scheduled via queue_sort_method and load_formula) while sending the others to PEs with other (appropriate) allocation rules that won't cause (ii). Well, I created an additional PE with alloacation_rule $pe_slots, and built in an if condition in pe.jsv for all jobs which request just a single node to be assigned to this new PE. But the annoying situation didn't change. The scheduler configuration is set to queue_sort_methodload and load_formula slots. So what I'm still missing? Ignore previous message. Me getting it back to front I think. That looks correct (I think). Have you checked the jobs show the right granted PE with qstat -j? Yes, of course. Sorry to step in the discussion: `qstat -j ...` shows the requested one, the granted one is in `qstat -r`. $ qsub -pe * 2 test.sh Your job 44329 (test.sh) has been submitted $ qstat -j 44329 ... parallel environment: * range: 2 ... $ qstat -r ... Requested PE: * 2 Granted PE: make 2 -- Reuti At the moment I don't know if I checked it with qstat -j, but I checked it - when I'm in the office again I probably have the output still on some screen window so I can tell it exactly. And I did do a test: I removed the PE temporarely from the queue - with the result that the jobs could not start anymore (as respected). ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
Hi Reuti, thank you for asking. The answer is unfortunately not. But at least the calculation is correct (qstat -g c), as soon as the user requests not all of the available slots. So the unsatisfying solution is that he's submitting his batch jobs now with less slots... Ursula -Ursprüngliche Nachricht- Von: Reuti [mailto:re...@staff.uni-marburg.de] Gesendet: Donnerstag, 23. Oktober 2014 11:26 An: Winkler, Ursula (ursula.wink...@uni-graz.at) Cc: William Hay; Gridengine Users Group Betreff: Re: AW: [gridengine users] Weird scheduler calculation error? Hi, Am 21.10.2014 um 14:23 schrieb Winkler, Ursula (ursula.wink...@uni-graz.at): schedd_job_info is switched on then? But even if switched off it should show up in `qalter -w p`. Yes, it is on. And 24 slots per machine then - `qstat -g c ` reveals the slots as being free too? A good question: it reveals that from 72 are now (!) 70 are free and no one is used. What's happening here? Did you get any further insight about this issue? -- Reuti Ursula And there is just that job and no other ones. Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
What is the setting of slots in the PE itself? 492(the cluster has 35 hosts with 12 slots and 3 hosts with 24 slots; but all hosts with 12 slots are disabled in this testqueue). Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
Hi Reuti, no - and there is no (other than reputed slots) resource shortage. And no host is in error state. Ursula -Ursprüngliche Nachricht- Von: Reuti [mailto:re...@staff.uni-marburg.de] Gesendet: Dienstag, 21. Oktober 2014 12:25 An: Gridengine Users Group Cc: Winkler, Ursula (ursula.wink...@uni-graz.at) Betreff: Re: [gridengine users] Weird scheduler calculation error? Hi, Am 21.10.2014 um 11:21 schrieb Ursula Winkler: Hi gridengine members, For now I ran out of ideas with an annoying problem: A job with 72 slots does not start because of qstat -j jobid tells cannot run in PE mpios because it only offers 64 slots, but there are 72 free (qalter -w p jobid and qalter -w -v jobid tells verification: found possible assignment with 72 slots and nothing more). The PE and host configurations are ok. I already restarted the master-process and the execution-daemons on the concerning hosts, but (as expected) that didn't help. Any calendar in effect? -- Reuti So what could be wrong? Any suggestions are appreciated. Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
The allocation rule is $fill_up and not $pe_slots. So that should be ok. Ursula -Ursprüngliche Nachricht- Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im Auftrag von Simon Andrews Gesendet: Dienstag, 21. Oktober 2014 13:05 An: Gridengine Users Group Betreff: Re: [gridengine users] Weird scheduler calculation error? What is your allocation rule for the mpios parallel environment (qconf -sp mpios)? Could it be that the allocation says that the slots all have to be on the same physical node, and no single node has more than 64 slots available? -Original Message- From: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] On Behalf Of Winkler, Ursula (ursula.wink...@uni-graz.at) Sent: 21 October 2014 11:46 To: Gridengine Users Group Subject: Re: [gridengine users] Weird scheduler calculation error? Hi Reuti, no - and there is no (other than reputed slots) resource shortage. And no host is in error state. Ursula -Ursprüngliche Nachricht- Von: Reuti [mailto:re...@staff.uni-marburg.de] Gesendet: Dienstag, 21. Oktober 2014 12:25 An: Gridengine Users Group Cc: Winkler, Ursula (ursula.wink...@uni-graz.at) Betreff: Re: [gridengine users] Weird scheduler calculation error? Hi, Am 21.10.2014 um 11:21 schrieb Ursula Winkler: Hi gridengine members, For now I ran out of ideas with an annoying problem: A job with 72 slots does not start because of qstat -j jobid tells cannot run in PE mpios because it only offers 64 slots, but there are 72 free (qalter -w p jobid and qalter -w -v jobid tells verification: found possible assignment with 72 slots and nothing more). The PE and host configurations are ok. I already restarted the master-process and the execution-daemons on the concerning hosts, but (as expected) that didn't help. Any calendar in effect? -- Reuti So what could be wrong? Any suggestions are appreciated. Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users The Babraham Institute, Babraham Research Campus, Cambridge CB22 3AT Registered Charity No. 1053902. The information transmitted in this email is directed only to the addressee. If you received this in error, please contact the sender and delete this email from your system. The contents of this e-mail are the views of the sender and do not necessarily represent the views of the Babraham Institute. Full conditions at: www.babraham.ac.ukhttp://www.babraham.ac.uk/terms ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
The qsub man page states that -w p and -w v don't take into account load values. Possibly the job is requesting a complex whose value is determined by a load sensor and the returned value is not suitable but not causing an alarm. Should not qstat -j jobid list the shortage of a complex? There's nothing (I already checked the job script of the user too). And the hosts on which the job should run are completely free. Alternatively perhaps a higher priority job with a resource reservation in place? We run with schedd_job_info false so I don't have much experience with it but I would expect qstat -j to take reservations into account while qalter -w [pv] can't as they're basically just doing the calculation for that one job. No. It's a testqueue with 3 hosts reserved only for that queue. And there is just that job and no other ones. Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Weird scheduler calculation error?
schedd_job_info is switched on then? But even if switched off it should show up in `qalter -w p`. Yes, it is on. And 24 slots per machine then - `qstat -g c ` reveals the slots as being free too? A good question: it reveals that from 72 are now (!) 70 are free and no one is used. What's happening here? Ursula And there is just that job and no other ones. Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] Hosts not fully used with fill up
Hi Rafa, for such purposes we have configured a separate Parallel Environment with allocation_rule $pe_slots (instead of $fill_up). Jobs scheduled with this rule can run ONLY on one host. Regards, Usula -Ursprüngliche Nachricht- Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im Auftrag von Rafael Arco Arredondo Gesendet: Donnerstag, 18. September 2014 09:44 An: users@gridengine.org Betreff: [gridengine users] Hosts not fully used with fill up Hello everyone, We are having an issue with the parallel environments and the allocation of slots with the fill up policy. Although we have configured the resource quotas of the queues not to use more than the number of slots the machine have and we control in the prolog that the jobs be submitted with a number of slots multiple of the number of physical processors, we are observing that sometimes, the slots of a job are split into several nodes, when they should be running in only one node. We are using Open Grid Scheduler 2011.11p1. This didn't happen in SGE 6.2. Has anyone experienced the same situation? Any clues of why it is happening? Thanks in advance, Rafa ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] submitting to a specific host
Hi, try qsub -q queue@host,... (e.g. qsub -q mpi@host1,mpi@host2, ) Ursula Von: users-boun...@gridengine.org [mailto:users-boun...@gridengine.org] Im Auftrag von Henrichs, Juryk Gesendet: Freitag, 21. März 2014 11:24 An: users@gridengine.org Betreff: [gridengine users] submitting to a specific host Hi, We are using SunGridEngine 6.5-2. To submit a job to a specific host on our cluster. I tried something like: qsub -l hostname=nodeXXX jobfile.sge Unfortunately the job does not get started although the nodeXXX is active in the requested queue and empty. qstat -j seems not to give a useful hint. However, the last two lines are: cannot run because it exceeds limit / in rule isis_licenses/1 cannot run in PE peSlots because it only offers 0 slots The resource isis_licenses is not requested here, so that shouldn't be a problem here? Any suggestions at where to start searching the problem? Juryk This e-mail and any attachment thereto may contain confidential information and/or information protected by intellectual property rights for the exclusive attention of the intended addressees named above. Any access of third parties to this e-mail is unauthorised. Any use of this e-mail by unintended recipients such as total or partial copying, distribution, disclosure etc. is prohibited and may be unlawful. When addressed to our clients the content of this e-mail is subject to the General Terms and Conditions of DNV GL Group of Companies applicable at the date of this e-mail. If you have received this e-mail in error, please notify the sender either by telephone or by e-mail and delete the material from any computer. DNV GL Group of Companies does not warrant and/or guarantee that this message at the moment of receipt is authentic, correct and its communication free of errors, interruption etc. FutureShip GmbH, HRB 106781 AG HH, VAT Reg. No. DE263937825 Geschäftsführer (CEO): Volker Höppner, Stefan Deucker ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
[gridengine users] qacct and resource requests
Dear gridengine list members, I've a question concerning qacct and resource requests. qacct prints out a lot of information - but little about resource requests that jobs asked for by qsub -l Of course it is possible to find out via qacct -l the jobs with a *particular* value (as one has to specify e.g. qacct -l h_vmem=2.5G ...) but there doesn't seem to exist a possibility to find out in a more direct way the values jobs have ordered (I need it for statistical purposes). Would be great if somebody has an idea. Cheers, Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] qacct and resource requests
I've a question concerning qacct and resource requests. qacct prints out a lot of information - but little about resource requests that jobs asked for by qsub -l Of course it is possible to find out via qacct -l the jobs with a *particular* value (as one has to specify e.g. qacct -l h_vmem=2.5G ...) but there doesn't seem to exist a possibility to find out in a more direct way the values jobs have ordered (I need it for statistical purposes). They are recorded in the accounting file in plain text in the field category (`man accounting`). -- Reuti Thank you, Reuti! Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
[gridengine users] Jobs being killed when another jobs start
Hi everyone, I have a problem with SGE: it looks like (though I'm really not sure) that running jobs are killed when other jobs are submitted which otherwise should wait in the queue because no free cores are available. It happens repeatedly so this raises suspicion. Unfortunately the error of the job log files only says that the jobs terminated badly. qacct -j job-no tells that the exit status of the jobs is 7. I could not find out what this error code means. Does anybody know that? Thanks, Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] qtcsh building error
Oh no, sorry, I was too hasty: adding -lcurses to gcc -o tcsh - ... -DPROG_NAME=qtcsh -lcrypt of course produced just tcsh, not qtcsh. Now, can I rename it to qtcsh, or what's about qtcsh? Ursula Von: Tim Wright [mailto:tim.dormo...@gmail.com] Gesendet: Mittwoch, 28. März 2012 17:22 An: Winkler, Ursula (ursula.wink...@uni-graz.at) Betreff: Re: [gridengine users] qtcsh building error Have you tried adding -lcurses or -lncurses to your compile string? On 28 March 2012 15:23, Ursula Winkler ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at wrote: undefined reference to `tputs' ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] qtcsh building error
Thank you! Ursula Von: Ron Chen [mailto:ron_chen_...@yahoo.com] Gesendet: Donnerstag, 29. März 2012 09:34 An: Winkler, Ursula (ursula.wink...@uni-graz.at); 'Tim Wright' Cc: 'users@gridengine.org' Betreff: Re: [gridengine users] qtcsh building error The install script copies tcsh to qtcsh into the final installable package. So you don't need to worry about not getting qtcsh. -Ron From: Winkler, Ursula (ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at) ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at To: 'Tim Wright' tim.dormo...@gmail.commailto:tim.dormo...@gmail.com Cc: 'users@gridengine.org' users@gridengine.orgmailto:users@gridengine.org Sent: Thursday, March 29, 2012 3:28 AM Subject: Re: [gridengine users] qtcsh building error Oh no, sorry, I was too hasty: adding “-lcurses” to “gcc –o tcsh - … -DPROG_NAME=”qtcsh” …. –lcrypt” of course produced just “tcsh”, not “qtcsh”. Now, can I rename it to qtcsh, or what’s about qtcsh? Ursula Von: Tim Wright [mailto:tim.dormo...@gmail.com] Gesendet: Mittwoch, 28. März 2012 17:22 An: Winkler, Ursula (ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at) Betreff: Re: [gridengine users] qtcsh building error Have you tried adding -lcurses or -lncurses to your compile string? On 28 March 2012 15:23, Ursula Winkler ursula.wink...@uni-graz.atmailto:ursula.wink...@uni-graz.at wrote: undefined reference to `tputs' ___ users mailing list users@gridengine.orgmailto:users@gridengine.org https://gridengine.org/mailman/listinfo/users ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
Re: [gridengine users] qtcsh building error
I know it's possible to build the binaries without qtcsh but I'm not sure if it is not needed later? Can I disclaim it? qtcsh is a version of tcsh where certain commands (which you can specify) are send to a node by SGE instead of being executed on the machine where you issued the command. If you don't use tcsh, you most likely won't need it. http://arc.liv.ac.uk/SGE/htmlman/htmlman1/qtcsh.html -- Reuti Then it's not so fine disclaiming it because I have users who use tcsh. Is there a workaround to get the compilation error solved? Ursula ___ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users