On Sat, Feb 06, 2021 at 04:35:48PM +0800, Lecopzer Chen wrote:
> Linux support KAsan for VMALLOC since commit 3c5c3cfb9ef4da9
> ("kasan: support backing vmalloc space with real shadow memory")
> 
> Like how the MODULES_VADDR does now, just not to early populate
> the VMALLOC_START between VMALLOC_END.
> 
> Before:
> 
> MODULE_VADDR: no mapping, no zoreo shadow at init
> VMALLOC_VADDR: backed with zero shadow at init
> 
> After:
> 
> MODULE_VADDR: no mapping, no zoreo shadow at init
> VMALLOC_VADDR: no mapping, no zoreo shadow at init

s/zoreo/zero/

> Thus the mapping will get allocated on demand by the core function
> of KASAN_VMALLOC.
> 
>   -----------  vmalloc_shadow_start
>  |           |
>  |           |
>  |           | <= non-mapping
>  |           |
>  |           |
>  |-----------|
>  |///////////|<- kimage shadow with page table mapping.
>  |-----------|
>  |           |
>  |           | <= non-mapping
>  |           |
>  ------------- vmalloc_shadow_end
>  |00000000000|
>  |00000000000| <= Zero shadow
>  |00000000000|
>  ------------- KASAN_SHADOW_END
> 
> Signed-off-by: Lecopzer Chen <lecopzer.c...@mediatek.com>
> ---
>  arch/arm64/mm/kasan_init.c | 18 +++++++++++++-----
>  1 file changed, 13 insertions(+), 5 deletions(-)
> 
> diff --git a/arch/arm64/mm/kasan_init.c b/arch/arm64/mm/kasan_init.c
> index d8e66c78440e..20d06008785f 100644
> --- a/arch/arm64/mm/kasan_init.c
> +++ b/arch/arm64/mm/kasan_init.c
> @@ -214,6 +214,7 @@ static void __init kasan_init_shadow(void)
>  {
>       u64 kimg_shadow_start, kimg_shadow_end;
>       u64 mod_shadow_start, mod_shadow_end;
> +     u64 vmalloc_shadow_end;
>       phys_addr_t pa_start, pa_end;
>       u64 i;
>  
> @@ -223,6 +224,8 @@ static void __init kasan_init_shadow(void)
>       mod_shadow_start = (u64)kasan_mem_to_shadow((void *)MODULES_VADDR);
>       mod_shadow_end = (u64)kasan_mem_to_shadow((void *)MODULES_END);
>  
> +     vmalloc_shadow_end = (u64)kasan_mem_to_shadow((void *)VMALLOC_END);
> +
>       /*
>        * We are going to perform proper setup of shadow memory.
>        * At first we should unmap early shadow (clear_pgds() call below).
> @@ -241,12 +244,17 @@ static void __init kasan_init_shadow(void)
>  
>       kasan_populate_early_shadow(kasan_mem_to_shadow((void *)PAGE_END),
>                                  (void *)mod_shadow_start);
> -     kasan_populate_early_shadow((void *)kimg_shadow_end,
> -                                (void *)KASAN_SHADOW_END);
>  
> -     if (kimg_shadow_start > mod_shadow_end)
> -             kasan_populate_early_shadow((void *)mod_shadow_end,
> -                                         (void *)kimg_shadow_start);

Not something introduced by this patch but what happens if this
condition is false? It means that kimg_shadow_end < mod_shadow_start and
the above kasan_populate_early_shadow(PAGE_END, mod_shadow_start)
overlaps with the earlier kasan_map_populate(kimg_shadow_start,
kimg_shadow_end).

> +     if (IS_ENABLED(CONFIG_KASAN_VMALLOC))
> +             kasan_populate_early_shadow((void *)vmalloc_shadow_end,
> +                                         (void *)KASAN_SHADOW_END);
> +     else {
> +             kasan_populate_early_shadow((void *)kimg_shadow_end,
> +                                         (void *)KASAN_SHADOW_END);
> +             if (kimg_shadow_start > mod_shadow_end)
> +                     kasan_populate_early_shadow((void *)mod_shadow_end,
> +                                                 (void *)kimg_shadow_start);
> +     }
>  
>       for_each_mem_range(i, &pa_start, &pa_end) {
>               void *start = (void *)__phys_to_virt(pa_start);
> -- 
> 2.25.1
> 

-- 
Catalin

Reply via email to