On Thu, Nov 28, 2013 at 10:55:06AM +0200, Gleb Natapov wrote:
> On Wed, Nov 27, 2013 at 09:06:36AM -0800, Paul E. McKenney wrote:
> > On Wed, Nov 27, 2013 at 10:00:09AM +0200, Gleb Natapov wrote:
> > > On Tue, Nov 26, 2013 at 11:35:06AM -0800, Paul E. McKenney wrote:
> > > > On Tue, Nov 26, 2013 at 06:24:13PM +0200, Michael S. Tsirkin wrote:
> > > > > On Wed, Sep 12, 2012 at 08:13:54AM -0700, Paul E. McKenney wrote:
> > > > > > On Wed, Sep 12, 2012 at 03:44:26PM +0300, Gleb Natapov wrote:
> > > > > > > On Wed, Sep 12, 2012 at 03:36:57PM +0300, Avi Kivity wrote:
> > > > > > > > On 09/12/2012 03:34 PM, Gleb Natapov wrote:
> > > > > > > > > On Wed, Sep 12, 2012 at 10:45:22AM +0300, Avi Kivity wrote:
> > > > > > > > >> On 09/12/2012 04:03 AM, Paul E. McKenney wrote:
> > > > > > > > >> >> > > Paul, I'd like to check something with you here:
> > > > > > > > >> >> > > this function can be triggered by userspace,
> > > > > > > > >> >> > > any number of times; we allocate
> > > > > > > > >> >> > > a 2K chunk of memory that is later freed by
> > > > > > > > >> >> > > kfree_rcu.
> > > > > > > > >> >> > >
> > > > > > > > >> >> > > Is there a risk of DOS if RCU is delayed while
> > > > > > > > >> >> > > lots of memory is queued up in this way?
> > > > > > > > >> >> > > If yes is this a generic problem with kfree_rcu
> > > > > > > > >> >> > > that should be addressed in core kernel?
> > > > > > > > >> >> >
> > > > > > > > >> >> > There is indeed a risk.
> > > > > > > > >> >>
> > > > > > > > >> >> In our case it's a 2K object. Is it a practical risk?
> > > > > > > > >> >
> > > > > > > > >> > How many kfree_rcu()s per second can a given user cause to
> > > > > > > > >> > happen?
> > > > > > > > >>
> > > > > > > > >> Not much more than a few hundred thousand per second per
> > > > > > > > >> process (normal
> > > > > > > > >> operation is zero).
> > > > > > > > >>
> > > > > > > > > I managed to do 21466 per second.
> > > > > > > >
> > > > > > > > Strange, why so slow?
> > > > > > > >
> > > > > > > Because ftrace buffer overflows :) With bigger buffer I get
> > > > > > > 169940.
> > > > > >
> > > > > > Ah, good, should not be a problem. In contrast, if you ran
> > > > > > kfree_rcu() in
> > > > > > a tight loop, you could probably do in excess of 100M per CPU per
> > > > > > second.
> > > > > > Now -that- might be a problem.
> > > > > >
> > > > > > Well, it -might- be a problem if you somehow figured out how to
> > > > > > allocate
> > > > > > memory that quickly in a steady-state manner. ;-)
> > > > > >
> > > > > > > > >> Good idea. Michael, is should be easy to modify
> > > > > > > > >> kvm-unit-tests to write
> > > > > > > > >> to the APIC ID register in a loop.
> > > > > > > > >>
> > > > > > > > > I did. Memory consumption does not grow on otherwise idle
> > > > > > > > > host.
> > > > > >
> > > > > > Very good -- the checks in __call_rcu(), which is common code
> > > > > > invoked by
> > > > > > kfree_rcu(), seem to be doing their job, then. These do keep a
> > > > > > per-CPU
> > > > > > counter, which can be adjusted via rcutree.blimit, which defaults
> > > > > > to taking evasive action if more than 10K callbacks are waiting on a
> > > > > > given CPU.
> > > > > >
> > > > > > My concern was that you might be overrunning that limit in way less
> > > > > > than a grace period (as in about a hundred microseconds. My concern
> > > > > > was of course unfounded -- you take several grace periods in push
> > > > > > 10K
> > > > > > callbacks through.
> > > > > >
> > > > > > Thanx, Paul
> > > > >
> > > > > Gleb noted that Documentation/RCU/checklist.txt has this text:
> > > > >
> > > > > An especially important property of the synchronize_rcu()
> > > > > primitive is that it automatically self-limits: if grace
> > > > > periods
> > > > > are delayed for whatever reason, then the synchronize_rcu()
> > > > > primitive will correspondingly delay updates. In contrast,
> > > > > code using call_rcu() should explicitly limit update rate in
> > > > > cases where grace periods are delayed, as failing to do so can
> > > > > result in excessive realtime latencies or even OOM conditions.
> > > > >
> > > > > If call_rcu is self-limiting maybe this should be documented ...
> > > >
> > > > It would be more accurate to say that takes has some measures to limit
> > > > the damage -- you can overwhelm these measures if you try hard enough.
> > > >
> > > The question is: Is it safe to have a call_rcu() without any additional
> > > rate limiting
> > > on user triggerable code path?
> >
> > That would be a good way to allow users to run your system out of memory,
> > especially on systems with limited memory. (If you have several GB of
> > free space, you might be OK.)
> >
> Thanks! Got it.
Does the following help?
Thanx, Paul
------------------------------------------------------------------------
rcu: Document call_rcu() safety mechanisms and limitations
The call_rcu() family of primitives will take action to accelerate
grace periods when the number of callbacks pending on a given CPU
becomes excessive. Although this safety mechanism can be useful,
it is no substitute for users of call_rcu() having rate-limit controls
in place. This commit adds this nuance to the documentation.
Reported-by: "Michael S. Tsirkin" <[email protected]>
Reported-by: Gleb Natapov <[email protected]>
Signed-off-by: Paul E. McKenney <[email protected]>
diff --git a/Documentation/RCU/checklist.txt b/Documentation/RCU/checklist.txt
index 91266193b8f4..5733e31836b5 100644
--- a/Documentation/RCU/checklist.txt
+++ b/Documentation/RCU/checklist.txt
@@ -256,10 +256,11 @@ over a rather long period of time, but improvements are
always welcome!
variations on this theme.
b. Limiting update rate. For example, if updates occur only
- once per hour, then no explicit rate limiting is required,
- unless your system is already badly broken. The dcache
- subsystem takes this approach -- updates are guarded
- by a global lock, limiting their rate.
+ once per hour, then no explicit rate limiting is
+ required, unless your system is already badly broken.
+ Older versions of the dcache subsystem takes this
+ approach -- updates were guarded by a global lock,
+ limiting their rate.
c. Trusted update -- if updates can only be done manually by
superuser or some other trusted user, then it might not
@@ -268,7 +269,8 @@ over a rather long period of time, but improvements are
always welcome!
the machine.
d. Use call_rcu_bh() rather than call_rcu(), in order to take
- advantage of call_rcu_bh()'s faster grace periods.
+ advantage of call_rcu_bh()'s faster grace periods. (This
+ is only a partial solution, though.)
e. Periodically invoke synchronize_rcu(), permitting a limited
number of updates per grace period.
@@ -276,6 +278,13 @@ over a rather long period of time, but improvements are
always welcome!
The same cautions apply to call_rcu_bh(), call_rcu_sched(),
call_srcu(), and kfree_rcu().
+ Note that although these primitives do take action to avoid memory
+ exhaustion when any given CPU has too many callbacks, a determined
+ user could still exhaust memory. This is especially the case
+ if a system with a large number of CPUs has been configured to
+ offload all of its RCU callbacks onto a single CPU, or if the
+ system has relatively little free memory.
+
9. All RCU list-traversal primitives, which include
rcu_dereference(), list_for_each_entry_rcu(), and
list_for_each_safe_rcu(), must be either within an RCU read-side
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html