TODO: no meaningful description here yet, please see the cover letter
      for this RFC series.

Signed-off-by: Andrey Konovalov <[email protected]>
Link: 
https://linux-review.googlesource.com/id/If7d37003875b2ed3e0935702c8015c223d6416a4
---
 mm/kasan/common.c  | 69 +++++++++++++++++++++++++---------------------
 mm/kasan/generic.c |  4 +++
 mm/kasan/hw_tags.c | 53 +++++++++++++++++++++++++++++++++++
 mm/kasan/kasan.h   |  8 ++++++
 mm/kasan/report.c  | 10 +++++--
 mm/kasan/sw_tags.c |  4 +++
 6 files changed, 115 insertions(+), 33 deletions(-)

diff --git a/mm/kasan/common.c b/mm/kasan/common.c
index a3e67d49b893..d642d5fce1e5 100644
--- a/mm/kasan/common.c
+++ b/mm/kasan/common.c
@@ -135,35 +135,37 @@ void kasan_cache_create(struct kmem_cache *cache, 
unsigned int *size,
        unsigned int redzone_size;
        int redzone_adjust;
 
-       /* Add alloc meta. */
-       cache->kasan_info.alloc_meta_offset = *size;
-       *size += sizeof(struct kasan_alloc_meta);
-
-       /* Add free meta. */
-       if (IS_ENABLED(CONFIG_KASAN_GENERIC) &&
-           (cache->flags & SLAB_TYPESAFE_BY_RCU || cache->ctor ||
-            cache->object_size < sizeof(struct kasan_free_meta))) {
-               cache->kasan_info.free_meta_offset = *size;
-               *size += sizeof(struct kasan_free_meta);
-       }
-
-       redzone_size = optimal_redzone(cache->object_size);
-       redzone_adjust = redzone_size - (*size - cache->object_size);
-       if (redzone_adjust > 0)
-               *size += redzone_adjust;
-
-       *size = min_t(unsigned int, KMALLOC_MAX_SIZE,
-                       max(*size, cache->object_size + redzone_size));
+       if (static_branch_unlikely(&kasan_debug)) {
+               /* Add alloc meta. */
+               cache->kasan_info.alloc_meta_offset = *size;
+               *size += sizeof(struct kasan_alloc_meta);
+
+               /* Add free meta. */
+               if (IS_ENABLED(CONFIG_KASAN_GENERIC) &&
+                   (cache->flags & SLAB_TYPESAFE_BY_RCU || cache->ctor ||
+                    cache->object_size < sizeof(struct kasan_free_meta))) {
+                       cache->kasan_info.free_meta_offset = *size;
+                       *size += sizeof(struct kasan_free_meta);
+               }
 
-       /*
-        * If the metadata doesn't fit, don't enable KASAN at all.
-        */
-       if (*size <= cache->kasan_info.alloc_meta_offset ||
-                       *size <= cache->kasan_info.free_meta_offset) {
-               cache->kasan_info.alloc_meta_offset = 0;
-               cache->kasan_info.free_meta_offset = 0;
-               *size = orig_size;
-               return;
+               redzone_size = optimal_redzone(cache->object_size);
+               redzone_adjust = redzone_size - (*size - cache->object_size);
+               if (redzone_adjust > 0)
+                       *size += redzone_adjust;
+
+               *size = min_t(unsigned int, KMALLOC_MAX_SIZE,
+                               max(*size, cache->object_size + redzone_size));
+
+               /*
+                * If the metadata doesn't fit, don't enable KASAN at all.
+                */
+               if (*size <= cache->kasan_info.alloc_meta_offset ||
+                               *size <= cache->kasan_info.free_meta_offset) {
+                       cache->kasan_info.alloc_meta_offset = 0;
+                       cache->kasan_info.free_meta_offset = 0;
+                       *size = orig_size;
+                       return;
+               }
        }
 
        *flags |= SLAB_KASAN;
@@ -180,6 +182,7 @@ size_t kasan_metadata_size(struct kmem_cache *cache)
 struct kasan_alloc_meta *kasan_get_alloc_meta(struct kmem_cache *cache,
                                              const void *object)
 {
+       WARN_ON(!static_branch_unlikely(&kasan_debug));
        return (void *)reset_tag(object) + cache->kasan_info.alloc_meta_offset;
 }
 
@@ -187,6 +190,7 @@ struct kasan_free_meta *kasan_get_free_meta(struct 
kmem_cache *cache,
                                            const void *object)
 {
        BUILD_BUG_ON(sizeof(struct kasan_free_meta) > 32);
+       WARN_ON(!static_branch_unlikely(&kasan_debug));
        return (void *)reset_tag(object) + cache->kasan_info.free_meta_offset;
 }
 
@@ -266,8 +270,10 @@ void * __must_check kasan_init_slab_obj(struct kmem_cache 
*cache,
        if (!(cache->flags & SLAB_KASAN))
                return (void *)object;
 
-       alloc_meta = kasan_get_alloc_meta(cache, object);
-       __memset(alloc_meta, 0, sizeof(*alloc_meta));
+       if (static_branch_unlikely(&kasan_debug)) {
+               alloc_meta = kasan_get_alloc_meta(cache, object);
+               __memset(alloc_meta, 0, sizeof(*alloc_meta));
+       }
 
        if (IS_ENABLED(CONFIG_KASAN_SW_TAGS) || 
IS_ENABLED(CONFIG_KASAN_HW_TAGS))
                object = set_tag(object, assign_tag(cache, object, true, 
false));
@@ -305,6 +311,7 @@ static bool __kasan_slab_free(struct kmem_cache *cache, 
void *object,
        kasan_poison_memory(object, rounded_up_size, KASAN_KMALLOC_FREE);
 
        if ((IS_ENABLED(CONFIG_KASAN_GENERIC) && !quarantine) ||
+                       !static_branch_unlikely(&kasan_debug) ||
                        unlikely(!(cache->flags & SLAB_KASAN)))
                return false;
 
@@ -351,7 +358,7 @@ static void *__kasan_kmalloc(struct kmem_cache *cache, 
const void *object,
        kasan_poison_memory((void *)redzone_start, redzone_end - redzone_start,
                KASAN_KMALLOC_REDZONE);
 
-       if (cache->flags & SLAB_KASAN)
+       if (static_branch_unlikely(&kasan_debug) && cache->flags & SLAB_KASAN)
                set_alloc_info(cache, (void *)object, flags);
 
        return set_tag(object, tag);
diff --git a/mm/kasan/generic.c b/mm/kasan/generic.c
index d259e4c3aefd..9d968eaedc98 100644
--- a/mm/kasan/generic.c
+++ b/mm/kasan/generic.c
@@ -33,6 +33,10 @@
 #include "kasan.h"
 #include "../slab.h"
 
+/* See the comments in hw_tags.c */
+DEFINE_STATIC_KEY_TRUE_RO(kasan_enabled);
+DEFINE_STATIC_KEY_TRUE_RO(kasan_debug);
+
 /*
  * All functions below always inlined so compiler could
  * perform better optimizations in each of __asan_loadX/__assn_storeX
diff --git a/mm/kasan/hw_tags.c b/mm/kasan/hw_tags.c
index b372421258c8..fc6ab1c8b155 100644
--- a/mm/kasan/hw_tags.c
+++ b/mm/kasan/hw_tags.c
@@ -8,6 +8,8 @@
 
 #define pr_fmt(fmt) "kasan: " fmt
 
+#include <linux/init.h>
+#include <linux/jump_label.h>
 #include <linux/kasan.h>
 #include <linux/kernel.h>
 #include <linux/memory.h>
@@ -17,8 +19,57 @@
 
 #include "kasan.h"
 
+enum kasan_mode {
+       KASAN_MODE_OFF,
+       KASAN_MODE_ON,
+       KASAN_MODE_DEBUG,
+};
+
+static enum kasan_mode kasan_mode __ro_after_init;
+
+/* Whether KASAN is enabled at all. */
+/* TODO: ideally no KASAN callbacks when this is disabled. */
+DEFINE_STATIC_KEY_FALSE_RO(kasan_enabled);
+
+/* Whether to collect debugging info, e.g. alloc/free stack traces. */
+DEFINE_STATIC_KEY_FALSE_RO(kasan_debug);
+
+/* Whether to use syncronous or asynchronous tag checking. */
+static bool kasan_sync __ro_after_init;
+
+static int __init early_kasan_mode(char *arg)
+{
+       if (!arg)
+               return -EINVAL;
+
+       if (strcmp(arg, "on") == 0)
+               kasan_mode = KASAN_MODE_ON;
+       else if (strcmp(arg, "debug") == 0)
+               kasan_mode = KASAN_MODE_DEBUG;
+       return 0;
+}
+early_param("kasan_mode", early_kasan_mode);
+
 void __init kasan_init_tags(void)
 {
+       /* TODO: system_supports_tags() always returns 0 here, fix. */
+       if (0 /*!system_supports_tags()*/)
+               return;
+
+       switch (kasan_mode) {
+       case KASAN_MODE_OFF:
+               return;
+       case KASAN_MODE_ON:
+               static_branch_enable(&kasan_enabled);
+               break;
+       case KASAN_MODE_DEBUG:
+               static_branch_enable(&kasan_enabled);
+               static_branch_enable(&kasan_debug);
+               kasan_sync = true;
+               break;
+       }
+
+       /* TODO: choose between sync and async based on kasan_sync. */
        init_tags(KASAN_TAG_MAX);
 
        pr_info("KernelAddressSanitizer initialized\n");
@@ -60,6 +111,7 @@ void kasan_set_free_info(struct kmem_cache *cache,
 {
        struct kasan_alloc_meta *alloc_meta;
 
+       WARN_ON(!static_branch_unlikely(&kasan_debug));
        alloc_meta = kasan_get_alloc_meta(cache, object);
        kasan_set_track(&alloc_meta->free_track[0], GFP_NOWAIT);
 }
@@ -69,6 +121,7 @@ struct kasan_track *kasan_get_free_track(struct kmem_cache 
*cache,
 {
        struct kasan_alloc_meta *alloc_meta;
 
+       WARN_ON(!static_branch_unlikely(&kasan_debug));
        alloc_meta = kasan_get_alloc_meta(cache, object);
        return &alloc_meta->free_track[0];
 }
diff --git a/mm/kasan/kasan.h b/mm/kasan/kasan.h
index 47d6074c7958..3712e7a39717 100644
--- a/mm/kasan/kasan.h
+++ b/mm/kasan/kasan.h
@@ -279,6 +279,14 @@ static inline const void *arch_kasan_set_tag(const void 
*addr, u8 tag)
 #define get_mem_tag(addr)                      arch_get_mem_tag(addr)
 #define set_mem_tag_range(addr, size, tag)     arch_set_mem_tag_range((addr), 
(size), (tag))
 
+#ifdef CONFIG_KASAN_HW_TAGS
+DECLARE_STATIC_KEY_FALSE(kasan_enabled);
+DECLARE_STATIC_KEY_FALSE(kasan_debug);
+#else
+DECLARE_STATIC_KEY_TRUE(kasan_enabled);
+DECLARE_STATIC_KEY_TRUE(kasan_debug);
+#endif
+
 /*
  * Exported functions for interfaces called from assembly or from generated
  * code. Declarations here to avoid warning about missing declarations.
diff --git a/mm/kasan/report.c b/mm/kasan/report.c
index dee5350b459c..ae956a29ad4e 100644
--- a/mm/kasan/report.c
+++ b/mm/kasan/report.c
@@ -159,8 +159,8 @@ static void describe_object_addr(struct kmem_cache *cache, 
void *object,
                (void *)(object_addr + cache->object_size));
 }
 
-static void describe_object(struct kmem_cache *cache, void *object,
-                               const void *addr, u8 tag)
+static void describe_object_stacks(struct kmem_cache *cache, void *object,
+                                       const void *addr, u8 tag)
 {
        struct kasan_alloc_meta *alloc_meta = kasan_get_alloc_meta(cache, 
object);
 
@@ -188,7 +188,13 @@ static void describe_object(struct kmem_cache *cache, void 
*object,
                }
 #endif
        }
+}
 
+static void describe_object(struct kmem_cache *cache, void *object,
+                               const void *addr, u8 tag)
+{
+       if (static_branch_unlikely(&kasan_debug))
+               describe_object_stacks(cache, object, addr, tag);
        describe_object_addr(cache, object, addr);
 }
 
diff --git a/mm/kasan/sw_tags.c b/mm/kasan/sw_tags.c
index 099af6dc8f7e..50e797a16e17 100644
--- a/mm/kasan/sw_tags.c
+++ b/mm/kasan/sw_tags.c
@@ -33,6 +33,10 @@
 #include "kasan.h"
 #include "../slab.h"
 
+/* See the comments in hw_tags.c */
+DEFINE_STATIC_KEY_TRUE_RO(kasan_enabled);
+DEFINE_STATIC_KEY_TRUE_RO(kasan_debug);
+
 static DEFINE_PER_CPU(u32, prng_state);
 
 void __init kasan_init_tags(void)
-- 
2.28.0.1011.ga647a8990f-goog

Reply via email to