Hi Kyrill,

On 2019/9/5 17:41, Kyrill Tkachov wrote:
> Hi Shaokun,
> 
> On 9/5/19 10:37 AM, Shaokun Zhang wrote:
>> Hi Kyrill,
>>
>> On 2019/9/3 19:13, Kyrill Tkachov wrote:
>>> Hi Shaokun,
>>>
>>> On 9/3/19 9:35 AM, Shaokun Zhang wrote:
>>>> The DCache clean & ICache invalidation requirements for instructions
>>>> to be data coherence are discoverable through new fields in CTR_EL0.
>>>> Let's support the two bits if they are enabled, the CPU core will
>>>> not execute the unnecessary DCache clean or Icache Invalidation
>>>> instructions.
>>>>
>>>> 2019-09-03  Shaokun Zhang <zhangshao...@hisilicon.com>
>>>>
>>>>      * config/aarch64/sync-cache.c: Support CTR_EL0.IDC and CTR_EL0.DIC in
>>>> __aarch64_sync_cache_range function.
>>>
>>> This version looks ok to me. I've checked the assembly generated for 
>>> sync-cache.c and it looks as expected.
>>>
>>> Can you please confirm that you've bootstrapped and tested this patch on an 
>>> aarch64 system?
>>>
>> Sorry to reply later.
>>
>> I have bootstrapped and also check the 
>> prev-aarch64-unknown-linux-gnu/libgcc/sync-cache.o file.
>> It has been compiled the patch.
>>
>> About the test, I run the self-modify testcase which calls __clear_cache 
>> directly,
>> and use the new GCC to compile it, it works ok on both Huawei Kunpeng 920 
>> server whose cpu
>> is HiSilicon tsv110 core and HiSilicon D05 server board which uses A72 core.
> 
> Thanks, that looks good to me now.
> 

Thanks your reply.

Shaokun

> Kyrill
> 
> 
>> Thanks,
>> Shaokun
>>
>>> Thanks,
>>>
>>> Kyrill
>>>
>>>
>>>> ---
>>>>   libgcc/config/aarch64/sync-cache.c | 57 
>>>> ++++++++++++++++++++++++--------------
>>>>   1 file changed, 36 insertions(+), 21 deletions(-)
>>>>
>>>> diff --git a/libgcc/config/aarch64/sync-cache.c 
>>>> b/libgcc/config/aarch64/sync-cache.c
>>>> index 791f5e42ff44..ea3da4be02b3 100644
>>>> --- a/libgcc/config/aarch64/sync-cache.c
>>>> +++ b/libgcc/config/aarch64/sync-cache.c
>>>> @@ -23,6 +23,9 @@ a copy of the GCC Runtime Library Exception along with 
>>>> this program;
>>>>   see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
>>>>   <http://www.gnu.org/licenses/>. */
>>>>
>>>> +#define CTR_IDC_SHIFT           28
>>>> +#define CTR_DIC_SHIFT           29
>>>> +
>>>>   void __aarch64_sync_cache_range (const void *, const void *);
>>>>
>>>>   void
>>>> @@ -41,32 +44,44 @@ __aarch64_sync_cache_range (const void *base, const 
>>>> void *end)
>>>>     icache_lsize = 4 << (cache_info & 0xF);
>>>>     dcache_lsize = 4 << ((cache_info >> 16) & 0xF);
>>>>
>>>> -  /* Loop over the address range, clearing one cache line at once.
>>>> -     Data cache must be flushed to unification first to make sure the
>>>> -     instruction cache fetches the updated data.  'end' is exclusive,
>>>> -     as per the GNU definition of __clear_cache.  */
>>>> +  /* If CTR_EL0.IDC is enabled, Data cache clean to the Point of 
>>>> Unification is
>>>> +     not required for instruction to data coherence.  */
>>>> +
>>>> +  if (((cache_info >> CTR_IDC_SHIFT) & 0x1) == 0x0) {
>>>> +    /* Loop over the address range, clearing one cache line at once.
>>>> +       Data cache must be flushed to unification first to make sure the
>>>> +       instruction cache fetches the updated data.  'end' is exclusive,
>>>> +       as per the GNU definition of __clear_cache.  */
>>>>
>>>> -  /* Make the start address of the loop cache aligned.  */
>>>> -  address = (const char*) ((__UINTPTR_TYPE__) base
>>>> -                          & ~ (__UINTPTR_TYPE__) (dcache_lsize - 1));
>>>> +    /* Make the start address of the loop cache aligned. */
>>>> +    address = (const char*) ((__UINTPTR_TYPE__) base
>>>> +                            & ~ (__UINTPTR_TYPE__) (dcache_lsize - 1));
>>>>
>>>> -  for (; address < (const char *) end; address += dcache_lsize)
>>>> -    asm volatile ("dc\tcvau, %0"
>>>> -                 :
>>>> -                 : "r" (address)
>>>> -                 : "memory");
>>>> +    for (; address < (const char *) end; address += dcache_lsize)
>>>> +      asm volatile ("dc\tcvau, %0"
>>>> +                   :
>>>> +                   : "r" (address)
>>>> +                   : "memory");
>>>> +  }
>>>>
>>>>     asm volatile ("dsb\tish" : : : "memory");
>>>>
>>>> -  /* Make the start address of the loop cache aligned.  */
>>>> -  address = (const char*) ((__UINTPTR_TYPE__) base
>>>> -                          & ~ (__UINTPTR_TYPE__) (icache_lsize - 1));
>>>> +  /* If CTR_EL0.DIC is enabled, Instruction cache cleaning to the Point of
>>>> +     Unification is not required for instruction to data coherence.  */
>>>> +
>>>> +  if (((cache_info >> CTR_DIC_SHIFT) & 0x1) == 0x0) {
>>>> +    /* Make the start address of the loop cache aligned. */
>>>> +    address = (const char*) ((__UINTPTR_TYPE__) base
>>>> +                            & ~ (__UINTPTR_TYPE__) (icache_lsize - 1));
>>>> +
>>>> +    for (; address < (const char *) end; address += icache_lsize)
>>>> +      asm volatile ("ic\tivau, %0"
>>>> +                   :
>>>> +                   : "r" (address)
>>>> +                   : "memory");
>>>>
>>>> -  for (; address < (const char *) end; address += icache_lsize)
>>>> -    asm volatile ("ic\tivau, %0"
>>>> -                 :
>>>> -                 : "r" (address)
>>>> -                 : "memory");
>>>> +    asm volatile ("dsb\tish" : : : "memory");
>>>> +  }
>>>>
>>>> -  asm volatile ("dsb\tish; isb" : : : "memory");
>>>> +  asm volatile("isb" : : : "memory");
>>>>   }
>>>> -- 
>>>> 2.7.4
>>>>
>>> .
>>>
> 
> .
> 

Reply via email to