I like both your suggestions.
I'll see if we can solve this "politically" before trying your approaches.
Thanks,
Txema
El 20/03/13 19:50, Reuti escribió:
Am 20.03.2013 um 18:21 schrieb Txema Heredia Genestar:
Well, right now my master node postfix has an alias mapping the user's ldap
email field. Thus, mails are not stored for the users but sent directly to
their real accounts.
Should I get rid of that, let all mail to accumulate in /var/spool/mail and
parse the files with a cron job?
Besides Joshua's idea:
You could use also a plain call to `mail`. I.e. if the there is local mail, it
can be retrieved by:
$ echo "p 1-10" | mail
already and an hourly cron-job could do:
$ echo "p (from root)" | mail | mail -s Summary [email protected]
(with the final address of the user). Or get the headers by:
$ mail -H
and you can do further processing.
-- Reuti
I have been looking for some parsers and they seem too complicated for what I
want to obtain. Is not there any mail client able to retrieve mails one by one
from a script?
Txema
El 20/03/13 13:50, Reuti escribió:
Hi,
Am 20.03.2013 um 13:01 schrieb Txema Heredia Genestar:
We have a 300-core cluster. Our users have always submitted their jobs using "-m
ea" to receive an email whenever the job finishes or aborts. Our typical user tends
to submit 1.000~3.000 jobs at once. They usually don't use task jobs, but submit each one
as an independent job. In addition, we have an epilog script that, whenever a job
finishes, checks the requested memory (h_vmem) and the real maxvmem used, and sends an
email to the user if they are blatantly requesting more than they need.
This leads that, on a typical "run", our users receive from 1.000 to 6.000
emails from the cluster.
Everybody was ok with this but, a few months ago, our university moved the mail
service to google and now the mail admins are threatening us to ban all mails
from our cluster.
Is there any solution to this (beyond not sending mails or changing the
task-job habits of our users)? Are there any addons/plugins out there that
accumulate mails in the master node and send hourly/daily digests to the users?
This was on the list before some times, to get only one email when the complete
array job finished in total. What you can implement:
a) Don't send emails from the array job, but submit a dummy job afterwards with
-hold_jid of the array job (it could even run in a dummy-queue on the headnode
solely for this purpose as it won't put any load on the machine), which sends
just one eMail after the main job. It could also perform a `qacct` to extract
the exit_codes from all the array instances.
b) Most likely you have an MTA agent running on the master node of the cluster
to forward it to Google right now. It should also be possible to use a
regexp/hash recipient_canonical to mangle the official email address like
[email protected] into foobar@headnode and collect them there first. A cron
job could then assemble one email per hour which includes all the emails for
this particular user of the last hour and send this one to Google.
-- Reuti
Thanks in advance,
Txema
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users