-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 On 15/01/14 22:28, Allison Walters wrote: > We have OpenMP jobs that need a user-defined (usually more than one > but less than all) number of cores on a single node for each job. > In addition to running these jobs, our program has an interface to > the cluster so they can submit jobs through a custom GUI (and we > build the qsub command in the background for the submission). I'm > trying to find a way for the job to request those multiple cores > that does not depend on the cluster to be configured a certain way, > since we have no control as to whether the client has a parallel > environment created, how it's named, etc... > > Basically, I'm just looking for the equivalent of -l nodes=[count] > in PBS/Torque, or -n [count] in LSF, etc... The program will use > the correct number of cores we pass to it, but we need to pass that > parameter to the cluster as well to ensure it only gets sent to a > node with the correct amount of cores available. This works fine > in the other clusters we support but I'm completely at a loss as to > how to do it in Grid Engine. I feel like I must be missing > something! :-) > > Thank you. I think the difference is underlying philosophy. Grid Engine:
1)Give a lot of flexibility to the administrator. This makes it easier for the admin to do things the designers didn't anticipate. It is less a batch scheduler more a batch scheduler construction kit with a reasonable sample configuration. Despite attempts to port it to Windows Grid Engine is basically a unix batch scheduler and follows the unix philosophy. 2)For the most part request the resources you need rather than try to tell the scheduler how to allocate jobs to nodes. Which resources you can request is up to the admin. Your actual problem isn't soluble in a configuration agnostic way but on many SGE clusters: There is a pe called smp with an allocation rule of $pe_slots. There is also commonly a resource called exclusive that can be requested to get exclusive access to a host. However this is just a common configuration and wouldn't work on the cluster I admin for example. You might choose to assume that anyone who doesn't use the configuration outlined above is familiar with their configuration and can adapt your job submission mechanism if you let them. If you want the more general ppn then the multiple PE's with allocation rules 1,2,3.. up to the largest number of cores on a single node should do it (never tried this just going by the man page). William -----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.11 (GNU/Linux) Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/ iQIcBAEBAgAGBQJS16S8AAoJEKCzH4joEjNW93UP/A6Nd0+wdKAENjsMKQfuoktL ZcGTUrMco7EaqV5V20UE1FHXfl26U/DM6tffPtWcQsFAGwRkYHNxEMCkE76F4CnQ M4x+IEJMMo4B85L5kp2/IiaoiE0Fe2AJCm8cez5kw1uEj39G+RVE0PDQW07HF8eG oksxHHKe+olr7BAEE5vdePhIjij78rH+H8qc6kr/KI507N0/eMocxkIKc26hC+JK zoTNnhgfC3VG4lnG1PUvqHXG9mYhZ+eaz+VLhHSaPnXoiZyP4KAREMfxX8DEnZSJ +IKibConmS8StWABBpmSxEfEZ/5rwqulvJP2UkrfwbZm8OzStNZjWnexCCoV9sJd GYtD66om4/3rMtnpPJYwemtII1qzo4OvZXEggCRPEsjBJTUitgeJ2RpfSqpyhMwO ABg4FY+//XDxr4WKrjpOy7dlEzCbCzadV5bh8aXhOILlVesbcZO1afVRKJ6vEil3 Zm0JasHL+pNLGDK1Qh5Zc37Gp9lsKgEK5ClbmSSzX+bjO2MeHkM7FUvzUHnqXo4c TIH95nM699X2CJju3ZIsSoO6yF3zRL8Yy1XCMkq3z7rNxx+jHq38bV+JZltyGNOe vfSyr0b3lAKDgdlAGnfYdlYj0KlThdWXQpMNUCrXagohIOrlT8LuJLJbOeLpVSS6 ztDhufhYTgUsclRefbBM =EuXr -----END PGP SIGNATURE----- _______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
