Makes Dept able to track dependencies by spinlock.

Signed-off-by: Byungchul Park <byungchul.p...@lge.com>
---
 include/linux/dept.h             |  4 +++-
 include/linux/llist.h            |  9 +--------
 include/linux/spinlock.h         | 42 ++++++++++++++++++++++++++++++++++++----
 include/linux/spinlock_api_smp.h | 15 ++++++++++++--
 include/linux/spinlock_types.h   | 37 ++++++++++++++++++++++++++++++-----
 include/linux/types.h            |  8 ++++++++
 kernel/locking/spinlock.c        |  6 +++++-
 kernel/locking/spinlock_debug.c  |  4 +++-
 kernel/sched/core.c              |  2 ++
 9 files changed, 105 insertions(+), 22 deletions(-)

diff --git a/include/linux/dept.h b/include/linux/dept.h
index 7fe1e04..baf3519 100644
--- a/include/linux/dept.h
+++ b/include/linux/dept.h
@@ -14,7 +14,9 @@
 #ifdef CONFIG_DEPT
 
 #include <linux/list.h>
-#include <linux/llist.h>
+#include <linux/types.h>
+
+struct task_struct;
 
 #define DEPT_MAX_STACK_ENTRY           16
 #define DEPT_MAX_WAIT_HIST             16
diff --git a/include/linux/llist.h b/include/linux/llist.h
index 2e9c721..e0c88d5 100644
--- a/include/linux/llist.h
+++ b/include/linux/llist.h
@@ -50,14 +50,7 @@
 
 #include <linux/atomic.h>
 #include <linux/kernel.h>
-
-struct llist_head {
-       struct llist_node *first;
-};
-
-struct llist_node {
-       struct llist_node *next;
-};
+#include <linux/types.h>
 
 #define LLIST_HEAD_INIT(name)  { NULL }
 #define LLIST_HEAD(name)       struct llist_head name = LLIST_HEAD_INIT(name)
diff --git a/include/linux/spinlock.h b/include/linux/spinlock.h
index f2f12d7..4c5c76e 100644
--- a/include/linux/spinlock.h
+++ b/include/linux/spinlock.h
@@ -92,15 +92,47 @@
 # include <linux/spinlock_up.h>
 #endif
 
+#ifdef CONFIG_DEPT
+#define dept_spin_init(m, k, s, n)             dept_map_init(m, k, s, n, 
DEPT_TYPE_SPIN)
+#define dept_spin_reinit(m, k, s, n)           dept_map_reinit(m, k, s, n)
+#define dept_spin_nocheck(m)                   dept_map_nocheck(m)
+#define dept_spin_lock(m, e_fn, ip)            dept_wait_ecxt_enter(m, 1UL, 
1UL, ip, __func__, __func__, e_fn, 0)
+#define dept_spin_lock_nest(m, n_m, ip)                WARN_ON(dept_top_map() 
!= (n_m))
+#define dept_spin_lock_nested(m, ne, e_fn, ip) dept_wait_ecxt_enter(m, 1UL, 
1UL, ip, __func__, __func__, e_fn, ne)
+#define dept_spin_trylock(m, e_fn, ip)         dept_ecxt_enter(m, 1UL, ip, 
__func__, e_fn, 0)
+#define dept_spin_unlock(m, ip)                        dept_ecxt_exit(m, ip)
+#define dept_spin_enter(m, e_fn, ip)           dept_ecxt_enter(m, 1UL, ip, 
"spin_lock_enter", e_fn, 0)
+#define dept_spin_exit(m, ip)                  dept_ecxt_exit(m, ip)
+#define dept_spin_switch_nested(m, ne, ip)                             \
+       do {                                                            \
+               dept_ecxt_exit(m, ip);                                  \
+               dept_ecxt_enter(m, 1UL, ip, __func__, "spin_switch_nested", 
ne);\
+       } while (0)
+#else
+#define dept_spin_init(m, k, s, n)             do { (void)(n); (void)(k); } 
while (0)
+#define dept_spin_reinit(m, k, s, n)           do { (void)(n); (void)(k); } 
while (0)
+#define dept_spin_nocheck(m)                   do { } while (0)
+#define dept_spin_lock(m, e_fn, ip)            do { } while (0)
+#define dept_spin_lock_nest(m, n_m, ip)                do { } while (0)
+#define dept_spin_lock_nested(m, ne, e_fn, ip) do { } while (0)
+#define dept_spin_trylock(m, e_fn, ip)         do { } while (0)
+#define dept_spin_unlock(m, ip)                        do { } while (0)
+#define dept_spin_enter(m, e_fn, ip)           do { } while (0)
+#define dept_spin_exit(m, ip)                  do { } while (0)
+#define dept_spin_switch_nested(m, ne, ip)     do { } while (0)
+#endif
+
 #ifdef CONFIG_DEBUG_SPINLOCK
   extern void __raw_spin_lock_init(raw_spinlock_t *lock, const char *name,
-                                  struct lock_class_key *key, short inner);
+                                  struct lock_class_key *key,
+                                  struct dept_key *dkey, short inner);
 
 # define raw_spin_lock_init(lock)                                      \
 do {                                                                   \
        static struct lock_class_key __key;                             \
+       static struct dept_key __dkey;                                  \
                                                                        \
-       __raw_spin_lock_init((lock), #lock, &__key, LD_WAIT_SPIN);      \
+       __raw_spin_lock_init((lock), #lock, &__key, &__dkey, LD_WAIT_SPIN); \
 } while (0)
 
 #else
@@ -231,7 +263,8 @@ static inline void do_raw_spin_unlock(raw_spinlock_t *lock) 
__releases(lock)
 # define raw_spin_lock_nest_lock(lock, nest_lock)                      \
         do {                                                           \
                 typecheck(struct lockdep_map *, &(nest_lock)->dep_map);\
-                _raw_spin_lock_nest_lock(lock, &(nest_lock)->dep_map); \
+                typecheck(struct dept_map *, &(nest_lock)->dmap);      \
+                _raw_spin_lock_nest_lock(lock, &(nest_lock)->dep_map, 
&(nest_lock)->dmap); \
         } while (0)
 #else
 /*
@@ -334,9 +367,10 @@ static __always_inline raw_spinlock_t 
*spinlock_check(spinlock_t *lock)
 # define spin_lock_init(lock)                                  \
 do {                                                           \
        static struct lock_class_key __key;                     \
+       static struct dept_key __dkey;                          \
                                                                \
        __raw_spin_lock_init(spinlock_check(lock),              \
-                            #lock, &__key, LD_WAIT_CONFIG);    \
+                            #lock, &__key, &__dkey, LD_WAIT_CONFIG); \
 } while (0)
 
 #else
diff --git a/include/linux/spinlock_api_smp.h b/include/linux/spinlock_api_smp.h
index 19a9be9..56afa5a 100644
--- a/include/linux/spinlock_api_smp.h
+++ b/include/linux/spinlock_api_smp.h
@@ -23,8 +23,8 @@
 void __lockfunc _raw_spin_lock_nested(raw_spinlock_t *lock, int subclass)
                                                                
__acquires(lock);
 void __lockfunc
-_raw_spin_lock_nest_lock(raw_spinlock_t *lock, struct lockdep_map *map)
-                                                               
__acquires(lock);
+_raw_spin_lock_nest_lock(raw_spinlock_t *lock, struct lockdep_map *map,
+               struct dept_map *dmap)                          
__acquires(lock);
 void __lockfunc _raw_spin_lock_bh(raw_spinlock_t *lock)                
__acquires(lock);
 void __lockfunc _raw_spin_lock_irq(raw_spinlock_t *lock)
                                                                
__acquires(lock);
@@ -88,6 +88,7 @@ static inline int __raw_spin_trylock(raw_spinlock_t *lock)
        preempt_disable();
        if (do_raw_spin_trylock(lock)) {
                spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
+               dept_spin_trylock(&lock->dmap, "__raw_spin_unlock", _RET_IP_);
                return 1;
        }
        preempt_enable();
@@ -108,6 +109,8 @@ static inline unsigned long 
__raw_spin_lock_irqsave(raw_spinlock_t *lock)
        local_irq_save(flags);
        preempt_disable();
        spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
+       dept_spin_lock(&lock->dmap, "__raw_spin_unlock_irqrestore", _RET_IP_);
+
        /*
         * On lockdep we dont want the hand-coded irq-enable of
         * do_raw_spin_lock_flags() code, because lockdep assumes
@@ -126,6 +129,7 @@ static inline void __raw_spin_lock_irq(raw_spinlock_t *lock)
        local_irq_disable();
        preempt_disable();
        spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
+       dept_spin_lock(&lock->dmap, "__raw_spin_unlock_irq", _RET_IP_);
        LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
 }
 
@@ -133,6 +137,7 @@ static inline void __raw_spin_lock_bh(raw_spinlock_t *lock)
 {
        __local_bh_disable_ip(_RET_IP_, SOFTIRQ_LOCK_OFFSET);
        spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
+       dept_spin_lock(&lock->dmap, "__raw_spin_unlock_bh", _RET_IP_);
        LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
 }
 
@@ -140,6 +145,7 @@ static inline void __raw_spin_lock(raw_spinlock_t *lock)
 {
        preempt_disable();
        spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
+       dept_spin_lock(&lock->dmap, "__raw_spin_unlock", _RET_IP_);
        LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
 }
 
@@ -148,6 +154,7 @@ static inline void __raw_spin_lock(raw_spinlock_t *lock)
 static inline void __raw_spin_unlock(raw_spinlock_t *lock)
 {
        spin_release(&lock->dep_map, _RET_IP_);
+       dept_spin_unlock(&lock->dmap, _RET_IP_);
        do_raw_spin_unlock(lock);
        preempt_enable();
 }
@@ -156,6 +163,7 @@ static inline void 
__raw_spin_unlock_irqrestore(raw_spinlock_t *lock,
                                            unsigned long flags)
 {
        spin_release(&lock->dep_map, _RET_IP_);
+       dept_spin_unlock(&lock->dmap, _RET_IP_);
        do_raw_spin_unlock(lock);
        local_irq_restore(flags);
        preempt_enable();
@@ -164,6 +172,7 @@ static inline void 
__raw_spin_unlock_irqrestore(raw_spinlock_t *lock,
 static inline void __raw_spin_unlock_irq(raw_spinlock_t *lock)
 {
        spin_release(&lock->dep_map, _RET_IP_);
+       dept_spin_unlock(&lock->dmap, _RET_IP_);
        do_raw_spin_unlock(lock);
        local_irq_enable();
        preempt_enable();
@@ -172,6 +181,7 @@ static inline void __raw_spin_unlock_irq(raw_spinlock_t 
*lock)
 static inline void __raw_spin_unlock_bh(raw_spinlock_t *lock)
 {
        spin_release(&lock->dep_map, _RET_IP_);
+       dept_spin_unlock(&lock->dmap, _RET_IP_);
        do_raw_spin_unlock(lock);
        __local_bh_enable_ip(_RET_IP_, SOFTIRQ_LOCK_OFFSET);
 }
@@ -181,6 +191,7 @@ static inline int __raw_spin_trylock_bh(raw_spinlock_t 
*lock)
        __local_bh_disable_ip(_RET_IP_, SOFTIRQ_LOCK_OFFSET);
        if (do_raw_spin_trylock(lock)) {
                spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
+               dept_spin_trylock(&lock->dmap, "__raw_spin_unlock_bh", 
_RET_IP_);
                return 1;
        }
        __local_bh_enable_ip(_RET_IP_, SOFTIRQ_LOCK_OFFSET);
diff --git a/include/linux/spinlock_types.h b/include/linux/spinlock_types.h
index b981caa..f180d71 100644
--- a/include/linux/spinlock_types.h
+++ b/include/linux/spinlock_types.h
@@ -16,6 +16,7 @@
 #endif
 
 #include <linux/lockdep_types.h>
+#include <linux/dept.h>
 
 typedef struct raw_spinlock {
        arch_spinlock_t raw_lock;
@@ -26,6 +27,7 @@
 #ifdef CONFIG_DEBUG_LOCK_ALLOC
        struct lockdep_map dep_map;
 #endif
+       struct dept_map dmap;
 } raw_spinlock_t;
 
 #define SPINLOCK_MAGIC         0xdead4ead
@@ -37,17 +39,33 @@
        .dep_map = {                                    \
                .name = #lockname,                      \
                .wait_type_inner = LD_WAIT_SPIN,        \
-       }
+       },
 # define SPIN_DEP_MAP_INIT(lockname)                   \
        .dep_map = {                                    \
                .name = #lockname,                      \
                .wait_type_inner = LD_WAIT_CONFIG,      \
-       }
+       },
 #else
 # define RAW_SPIN_DEP_MAP_INIT(lockname)
 # define SPIN_DEP_MAP_INIT(lockname)
 #endif
 
+#ifdef CONFIG_DEPT
+# define RAW_SPIN_DMAP_INIT(lockname)                  \
+       .dmap = {                                       \
+               .name = #lockname,                      \
+               .type = DEPT_TYPE_SPIN,                 \
+       },
+# define SPIN_DMAP_INIT(lockname)                      \
+       .dmap = {                                       \
+               .name = #lockname,                      \
+               .type = DEPT_TYPE_SPIN,                 \
+       },
+#else
+# define RAW_SPIN_DMAP_INIT(lockname)
+# define SPIN_DMAP_INIT(lockname)
+#endif
+
 #ifdef CONFIG_DEBUG_SPINLOCK
 # define SPIN_DEBUG_INIT(lockname)             \
        .magic = SPINLOCK_MAGIC,                \
@@ -61,7 +79,8 @@
        {                                       \
        .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED,  \
        SPIN_DEBUG_INIT(lockname)               \
-       RAW_SPIN_DEP_MAP_INIT(lockname) }
+       RAW_SPIN_DEP_MAP_INIT(lockname)         \
+       RAW_SPIN_DMAP_INIT(lockname) }
 
 #define __RAW_SPIN_LOCK_UNLOCKED(lockname)     \
        (raw_spinlock_t) __RAW_SPIN_LOCK_INITIALIZER(lockname)
@@ -73,11 +92,18 @@
                struct raw_spinlock rlock;
 
 #ifdef CONFIG_DEBUG_LOCK_ALLOC
-# define LOCK_PADSIZE (offsetof(struct raw_spinlock, dep_map))
+#define LOCK_PADSIZE (offsetof(struct raw_spinlock, dep_map))
                struct {
                        u8 __padding[LOCK_PADSIZE];
                        struct lockdep_map dep_map;
                };
+#undef LOCK_PADSIZE
+#define LOCK_PADSIZE (offsetof(struct raw_spinlock, dmap))
+               struct {
+                       u8 __padding_for_dept[LOCK_PADSIZE];
+                       struct dept_map dmap;
+               };
+#undef LOCK_PADSIZE
 #endif
        };
 } spinlock_t;
@@ -86,7 +112,8 @@
        {                                       \
        .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED,  \
        SPIN_DEBUG_INIT(lockname)               \
-       SPIN_DEP_MAP_INIT(lockname) }
+       SPIN_DEP_MAP_INIT(lockname)             \
+       SPIN_DMAP_INIT(lockname) }
 
 #define __SPIN_LOCK_INITIALIZER(lockname) \
        { { .rlock = ___SPIN_LOCK_INITIALIZER(lockname) } }
diff --git a/include/linux/types.h b/include/linux/types.h
index a147977..0b058f5 100644
--- a/include/linux/types.h
+++ b/include/linux/types.h
@@ -187,6 +187,14 @@ struct hlist_node {
        struct hlist_node *next, **pprev;
 };
 
+struct llist_head {
+       struct llist_node *first;
+};
+
+struct llist_node {
+       struct llist_node *next;
+};
+
 struct ustat {
        __kernel_daddr_t        f_tfree;
        __kernel_ino_t          f_tinode;
diff --git a/kernel/locking/spinlock.c b/kernel/locking/spinlock.c
index 0ff0838..e838e0e 100644
--- a/kernel/locking/spinlock.c
+++ b/kernel/locking/spinlock.c
@@ -359,6 +359,7 @@ void __lockfunc _raw_spin_lock_nested(raw_spinlock_t *lock, 
int subclass)
 {
        preempt_disable();
        spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_);
+       dept_spin_lock_nested(&lock->dmap, subclass, "_raw_spin_unlock", 
_RET_IP_);
        LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
 }
 EXPORT_SYMBOL(_raw_spin_lock_nested);
@@ -371,6 +372,7 @@ unsigned long __lockfunc 
_raw_spin_lock_irqsave_nested(raw_spinlock_t *lock,
        local_irq_save(flags);
        preempt_disable();
        spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_);
+       dept_spin_lock_nested(&lock->dmap, subclass, 
"_raw_spin_unlock_irqrestore", _RET_IP_);
        LOCK_CONTENDED_FLAGS(lock, do_raw_spin_trylock, do_raw_spin_lock,
                                do_raw_spin_lock_flags, &flags);
        return flags;
@@ -378,10 +380,12 @@ unsigned long __lockfunc 
_raw_spin_lock_irqsave_nested(raw_spinlock_t *lock,
 EXPORT_SYMBOL(_raw_spin_lock_irqsave_nested);
 
 void __lockfunc _raw_spin_lock_nest_lock(raw_spinlock_t *lock,
-                                    struct lockdep_map *nest_lock)
+                                    struct lockdep_map *nest_lock,
+                                    struct dept_map *nest_dmap)
 {
        preempt_disable();
        spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_);
+       dept_spin_lock_nest(&lock->dmap, nest_dmap, _RET_IP_);
        LOCK_CONTENDED(lock, do_raw_spin_trylock, do_raw_spin_lock);
 }
 EXPORT_SYMBOL(_raw_spin_lock_nest_lock);
diff --git a/kernel/locking/spinlock_debug.c b/kernel/locking/spinlock_debug.c
index b9d9308..03e6812 100644
--- a/kernel/locking/spinlock_debug.c
+++ b/kernel/locking/spinlock_debug.c
@@ -14,7 +14,8 @@
 #include <linux/export.h>
 
 void __raw_spin_lock_init(raw_spinlock_t *lock, const char *name,
-                         struct lock_class_key *key, short inner)
+                         struct lock_class_key *key,
+                         struct dept_key *dkey, short inner)
 {
 #ifdef CONFIG_DEBUG_LOCK_ALLOC
        /*
@@ -22,6 +23,7 @@ void __raw_spin_lock_init(raw_spinlock_t *lock, const char 
*name,
         */
        debug_check_no_locks_freed((void *)lock, sizeof(*lock));
        lockdep_init_map_wait(&lock->dep_map, name, key, 0, inner);
+       dept_spin_init(&lock->dmap, dkey, 0, name);
 #endif
        lock->raw_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
        lock->magic = SPINLOCK_MAGIC;
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 2d95dc3..f11ef3d 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -3500,6 +3500,7 @@ static inline void finish_task(struct task_struct *prev)
         */
        rq_unpin_lock(rq, rf);
        spin_release(&rq->lock.dep_map, _THIS_IP_);
+       dept_spin_exit(&rq->lock.dmap, _THIS_IP_);
 #ifdef CONFIG_DEBUG_SPINLOCK
        /* this is a valid case when another task releases the spinlock */
        rq->lock.owner = next;
@@ -3514,6 +3515,7 @@ static inline void finish_lock_switch(struct rq *rq)
         * prev into current:
         */
        spin_acquire(&rq->lock.dep_map, 0, 0, _THIS_IP_);
+       dept_spin_enter(&rq->lock.dmap, "raw_spin_inlock_irq", _THIS_IP_);
        raw_spin_unlock_irq(&rq->lock);
 }
 
-- 
1.9.1

Reply via email to