On Thu, Jan 14, 2021 at 08:36PM +0100, Andrey Konovalov wrote:
> Generic mm functions that call KASAN annotations that might report a bug
> pass _RET_IP_ to them as an argument. This allows KASAN to include the
> name of the function that called the mm function in its report's header.
> 
> Now that KASAN has inline wrappers for all of its annotations, move
> _RET_IP_ to those wrappers to simplify annotation call sites.
> 
> Link: 
> https://linux-review.googlesource.com/id/I8fb3c06d49671305ee184175a39591bc26647a67
> Signed-off-by: Andrey Konovalov <[email protected]>

Much nicer!

Reviewed-by: Marco Elver <[email protected]>

> ---
>  include/linux/kasan.h | 20 +++++++++-----------
>  mm/mempool.c          |  2 +-
>  mm/slab.c             |  2 +-
>  mm/slub.c             |  4 ++--
>  4 files changed, 13 insertions(+), 15 deletions(-)
> 
> diff --git a/include/linux/kasan.h b/include/linux/kasan.h
> index 5e0655fb2a6f..bba1637827c3 100644
> --- a/include/linux/kasan.h
> +++ b/include/linux/kasan.h
> @@ -181,19 +181,18 @@ static __always_inline void * __must_check 
> kasan_init_slab_obj(
>  }
>  
>  bool __kasan_slab_free(struct kmem_cache *s, void *object, unsigned long ip);
> -static __always_inline bool kasan_slab_free(struct kmem_cache *s, void 
> *object,
> -                                             unsigned long ip)
> +static __always_inline bool kasan_slab_free(struct kmem_cache *s, void 
> *object)
>  {
>       if (kasan_enabled())
> -             return __kasan_slab_free(s, object, ip);
> +             return __kasan_slab_free(s, object, _RET_IP_);
>       return false;
>  }
>  
>  void __kasan_slab_free_mempool(void *ptr, unsigned long ip);
> -static __always_inline void kasan_slab_free_mempool(void *ptr, unsigned long 
> ip)
> +static __always_inline void kasan_slab_free_mempool(void *ptr)
>  {
>       if (kasan_enabled())
> -             __kasan_slab_free_mempool(ptr, ip);
> +             __kasan_slab_free_mempool(ptr, _RET_IP_);
>  }
>  
>  void * __must_check __kasan_slab_alloc(struct kmem_cache *s,
> @@ -237,10 +236,10 @@ static __always_inline void * __must_check 
> kasan_krealloc(const void *object,
>  }
>  
>  void __kasan_kfree_large(void *ptr, unsigned long ip);
> -static __always_inline void kasan_kfree_large(void *ptr, unsigned long ip)
> +static __always_inline void kasan_kfree_large(void *ptr)
>  {
>       if (kasan_enabled())
> -             __kasan_kfree_large(ptr, ip);
> +             __kasan_kfree_large(ptr, _RET_IP_);
>  }
>  
>  bool kasan_save_enable_multi_shot(void);
> @@ -273,12 +272,11 @@ static inline void *kasan_init_slab_obj(struct 
> kmem_cache *cache,
>  {
>       return (void *)object;
>  }
> -static inline bool kasan_slab_free(struct kmem_cache *s, void *object,
> -                                unsigned long ip)
> +static inline bool kasan_slab_free(struct kmem_cache *s, void *object)
>  {
>       return false;
>  }
> -static inline void kasan_slab_free_mempool(void *ptr, unsigned long ip) {}
> +static inline void kasan_slab_free_mempool(void *ptr) {}
>  static inline void *kasan_slab_alloc(struct kmem_cache *s, void *object,
>                                  gfp_t flags)
>  {
> @@ -298,7 +296,7 @@ static inline void *kasan_krealloc(const void *object, 
> size_t new_size,
>  {
>       return (void *)object;
>  }
> -static inline void kasan_kfree_large(void *ptr, unsigned long ip) {}
> +static inline void kasan_kfree_large(void *ptr) {}
>  
>  #endif /* CONFIG_KASAN */
>  
> diff --git a/mm/mempool.c b/mm/mempool.c
> index 624ed51b060f..79959fac27d7 100644
> --- a/mm/mempool.c
> +++ b/mm/mempool.c
> @@ -104,7 +104,7 @@ static inline void poison_element(mempool_t *pool, void 
> *element)
>  static __always_inline void kasan_poison_element(mempool_t *pool, void 
> *element)
>  {
>       if (pool->alloc == mempool_alloc_slab || pool->alloc == mempool_kmalloc)
> -             kasan_slab_free_mempool(element, _RET_IP_);
> +             kasan_slab_free_mempool(element);
>       else if (pool->alloc == mempool_alloc_pages)
>               kasan_free_pages(element, (unsigned long)pool->pool_data);
>  }
> diff --git a/mm/slab.c b/mm/slab.c
> index d7c8da9319c7..afeb6191fb1e 100644
> --- a/mm/slab.c
> +++ b/mm/slab.c
> @@ -3421,7 +3421,7 @@ static __always_inline void __cache_free(struct 
> kmem_cache *cachep, void *objp,
>               memset(objp, 0, cachep->object_size);
>  
>       /* Put the object into the quarantine, don't touch it for now. */
> -     if (kasan_slab_free(cachep, objp, _RET_IP_))
> +     if (kasan_slab_free(cachep, objp))
>               return;
>  
>       /* Use KCSAN to help debug racy use-after-free. */
> diff --git a/mm/slub.c b/mm/slub.c
> index 75fb097d990d..0afb53488238 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -1514,7 +1514,7 @@ static inline void *kmalloc_large_node_hook(void *ptr, 
> size_t size, gfp_t flags)
>  static __always_inline void kfree_hook(void *x)
>  {
>       kmemleak_free(x);
> -     kasan_kfree_large(x, _RET_IP_);
> +     kasan_kfree_large(x);
>  }
>  
>  static __always_inline bool slab_free_hook(struct kmem_cache *s, void *x)
> @@ -1544,7 +1544,7 @@ static __always_inline bool slab_free_hook(struct 
> kmem_cache *s, void *x)
>                                    KCSAN_ACCESS_WRITE | KCSAN_ACCESS_ASSERT);
>  
>       /* KASAN might put x into memory quarantine, delaying its reuse */
> -     return kasan_slab_free(s, x, _RET_IP_);
> +     return kasan_slab_free(s, x);
>  }
>  
>  static inline bool slab_free_freelist_hook(struct kmem_cache *s,
> -- 
> 2.30.0.284.gd98b1dd5eaa7-goog
> 

Reply via email to