I believe you're wrong. The max_jobs_per_day criteria is applied to all 
host/app versions. The only relationship to quota is trust requires at least 10 
consecutive valid which means a higher quota than a new host or one which has 
been recently punished.
-- 
                                                      Joe

On Wed, 12 Jan 2011 13:52:47 -0500, <[email protected]> wrote:

>
>    The trusted computers would not have any quotas applied at all.
>    jm7
>    -----<[email protected]> wrote: -----
>
>      To: "Josef W. Segur" <[email protected]>, <[email protected]>
>      From: Raistmer <[email protected]>
>      Sent by: <[email protected]>
>      Date: 01/07/2011 04:22PM
>      cc: <[email protected]>, <[email protected]>
>      Subject: Re: [boinc_dev] Quota system inefficiency, something better?
>      This implies same quota mechanism as used today but just with different
>      numbers.
>      For GPU idle time will be long enough to make recovering from single act
>      of task trashing (it happens sometimes) very painful.
>      Maybe better to slightly change quota system to take into account history
>      of host behavior: what % of invalids this host has? And base speed of
>      quota recovery on this percent.
>      There  was  some  idea of "trusted" hosts in regards not to do task
>      replication if it is sent to "trusted" host.
>      Though I think it's bad idea, "trusted host" conception can be still used
>      for quota-related calculations IMHO.
>      ----- Original Message -----
>      From: [email protected]
>      To: Josef W. Segur
>      Cc: [email protected] ; [email protected]
>      Sent: Friday, January 07, 2011 10:22 PM
>      Subject: Re: [boinc_dev] Quota system inefficiency, something better?
>      I believe that doubling is way too fast.  My thought is that when you
>      return a good one, you should get a replacement and your quota should go
>      up
>      by one.  Yes, the recovery will be slower, but tomorrow you will be able
>      to
>      fetch  the  count of successful ones that you returned today at the
>      beginning
>      of the day.  If you are always on, you will still be able to keep your
>      CPUs
>      busy.  If you are recovering from a temporary problem, either you can
>      nursemaid your computer through a couple of days or you can just accept a
>      slightly idle CPU for a while if you are not always attached to the
>      internet.
>      jm7
>
>                  "Josef W. Segur"
>                  <jsegur@westelcom
>                  .com>                                                      To
>                  Sent by:                  <[email protected]>
>                  <boinc_dev-bounce                                          cc
>                  [email protected]
>                  u>                                                    Subject
>                                            Re: [boinc_dev] Quota system
>                                            inefficiency, something better?
>                  01/07/2011 02:04
>                  PM
>
>
>
>
>      Although the quota system is not limiting those hosts enough,
>      they get frequent invalidations. The consecutive valid count is
>      usually zero, though I've seen a few cases of single digit
>      non-zero counts. So with S@H gpu_multiplier set at 8, those hosts
>      are limited to not much more than 800 tasks per day per FERMI
>      GPU. My most recent check shows 19 hosts, though a few have two
>      Fermi cards so the GPU count is about 23. So the size of the
>      problem is reduced to under 2% of S@H Enhanced tasks.
>      S@H Enhanced tasks account for about 64% of the download
>      bandwidth, the much larger Astropulse v505 tasks get the
>      remainder. That makes the bandwidth impact on the order of 1%
>      currently.
>      Perhaps something like keeping a consecutive invalid count and
>      not doubling on reported "Success" when that count exceeds the
>      consecutive valid count would be better. But that would make
>      recovery from a temporary problem much slower.
>      --
>                                                                   Joe
>      On Fri, 07 Jan 2011 06:41:07 -0500, Richard Haselgrove wrote:
>      >  OK, back-of-envelope error - that 10% overstates the size of the
>      problem.
>      >
>      > But, more realistically, 20 hosts @ 2,000 tasks wasted each per day is
>      close to 4% of all multibeam tasks issued.
>      >
>      >   ----- Original Message -----
>      >   From: Richard Haselgrove
>      >
>      >
>      >   There is an updated version of that list at
>      
> [1]http://setiathome.berkeley.edu/forum_thread.php?id=62573&nowrap=true#10
>      62822
>      >
>      >   Given the limited number of hosts involved, in the short term the
>      wastage (something of the order of 10% of SETI's limited download
>      bandwidth), maybe could/should be stemmed by invoking
>      [2]http://boinc.berkeley.edu/trac/wiki/BlackListfor  the  19  hosts
>      identified.
>      >
>      >   This approach has been used successfully by Milo at CPDN, although
>      there is a suggestion that something (perhaps the replacement of
>      max_results_day with per_app_version equivalents) broke the blacklist
>      facility after he started using it - quotas which had been set to -1
>      manually became positive again. The SETI situation would be a useful test
>      of this tool while a longer-term automatic solution is sought.
>      >
>      >     ----- Original Message -----
>      >     From: Raistmer
>      >
>      >
>      >     Hello.
>      >
>      >     Looks like current quota system implementation can't prevent 
> project
>      resources waste in case of "partially" broken host.
>      >     For example, host with anonymous platform running FERMI 
> incompatible
>      CUDA app on SETI.
>      >     It will produce incorrect overflows almost always but few specific
>      ARs that will be processed correctly and recive validation.
>      >     This small nomber of validations + "GPU" status of app (GPU has
>      greatly relaxed limits) allows continuous task trashing. Current quota
>      system implementation can't prevent massive task trashing in this
>      situation.
>      >
>      >     But now more historical info about host behavior stored on servers,
>      on per app version basis.
>      >     Maybe smth new can be implemented that will take into account not
>      only last successive validation but host history too?
>      >        The  testcases  are known, SETI community has list of such
>      bad-behaving
>      hosts already:
>      
> [3]http://setiathome.berkeley.edu/forum_thread.php?id=62573&nowrap=true#10
>      61788
>      >
>      >     The aim should be to reduce their throughput to 1 task per day for
>      NV
>      GPU app until their owners reinstall GPU app.
_______________________________________________
boinc_dev mailing list
[email protected]
http://lists.ssl.berkeley.edu/mailman/listinfo/boinc_dev
To unsubscribe, visit the above URL and
(near bottom of page) enter your email address.

Reply via email to