In order to avoid races between setting a struct clock_event_device's
->mult_adjusted in clockevents_update_freq() and yet to be implemented
updates triggered from the timekeeping core, the setting of ->mult and
->mult_adjusted should be made atomic.

Protect the update in clockevents_update_freq() by locking the
clockevents_lock spinlock. Frequency updates are expected to be done
seldomly and thus, taking this subsystem lock should not have any impact
on performance.

Note that the call to tick_broadcast_update_freq() is also protected
by this clockevents_lock and thus, this patch can take the
tick_broadcast_lock with the clockevents_lock being held. However,
this is supposed to be safe since the sequence
  __clockevents_unbind() -> clockevents_replace() ->
  tick_install_replacement() -> tick_setup_device() ->
  tick_device_uses_broadcast()
already does this, too.

Signed-off-by: Nicolai Stange <nicsta...@gmail.com>
---
 kernel/time/clockevents.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/kernel/time/clockevents.c b/kernel/time/clockevents.c
index 67d572e..b121932 100644
--- a/kernel/time/clockevents.c
+++ b/kernel/time/clockevents.c
@@ -658,11 +658,11 @@ int clockevents_update_freq(struct clock_event_device 
*dev, u32 freq)
        unsigned long flags;
        int ret;
 
-       local_irq_save(flags);
+       raw_spin_lock_irqsave(&clockevents_lock, flags);
        ret = tick_broadcast_update_freq(dev, freq);
        if (ret == -ENODEV)
                ret = __clockevents_update_freq(dev, freq);
-       local_irq_restore(flags);
+       raw_spin_unlock_irqrestore(&clockevents_lock, flags);
        return ret;
 }
 
-- 
2.10.0

Reply via email to